Transitive hashing network for heterogeneous multimedia retrieval

Zhangjie Cao, Mingsheng Long*, Jianmin Wang, Qiang Yang

*Corresponding author for this work

Research output: Contribution to conferenceConference Paperpeer-review

25 Citations (Scopus)

Abstract

Hashing is widely applied to large-scale multimedia retrieval due to the storage and retrieval efficiency. Crossmodal hashing enables efficient retrieval of one modality from database relevant to a query of another modality. Existing work on cross-modal hashing assumes that heterogeneous relationship across modalities is available for learning to hash. This paper relaxes this strict assumption by only requiring heterogeneous relationship in some auxiliary dataset different from the query or database domain. We design a novel hybrid deep architecture, transitive hashing network (THN), to jointly learn cross-modal correlation from the auxiliary dataset, and align the data distributions of the auxiliary dataset with that of the query or database domain, which generates compact transitive hash codes for efficient crossmodal retrieval. Comprehensive empirical evidence validates that the proposed THN approach yields state of the art retrieval performance on standard multimedia benchmarks, i.e. NUS-WIDE and ImageNet-YahooQA.

Original languageEnglish
Pages81-87
Number of pages7
Publication statusPublished - 2017
Event31st AAAI Conference on Artificial Intelligence, AAAI 2017 - San Francisco, United States
Duration: 4 Feb 201710 Feb 2017

Conference

Conference31st AAAI Conference on Artificial Intelligence, AAAI 2017
Country/TerritoryUnited States
CitySan Francisco
Period4/02/1710/02/17

Bibliographical note

Publisher Copyright:
Copyright © 2017, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.

Fingerprint

Dive into the research topics of 'Transitive hashing network for heterogeneous multimedia retrieval'. Together they form a unique fingerprint.

Cite this