Cross-Modal Hashing Through Ranking Subspace Learning
Keywords
crossmodal hashing; Multimedia retrieval; ranking subspace learning; WTA hash
Abstract
Hashing has been widely used for approximate nearest neighbor search of high-dimensional multimedia data. In this paper, we propose a novel hash learning framework that maps high-dimensional multimodal data into a common Hamming space where the cross-modal similarity can be measured using Hamming distance. Unlike existing cross-modal hashing methods that learn hash functions in the form of numeric quantization of linear projections, the proposed hash learning algorithm encodes features' ranking properties and takes advantage of rank correlations which are known to be scale-invariant, numerically stable and highly nonlinear. Specifically, we learn two groups of subspaces jointly, one for each modality, so that the ranking orders in those subspaces maximally preserve the cross-modal similarity. Extensive experiments on realworld datasets demonstrate superiority of the proposed methods compared to state-of-the-arts.
Publication Date
8-25-2016
Publication Title
Proceedings - IEEE International Conference on Multimedia and Expo
Volume
2016-August
Document Type
Article; Proceedings Paper
Personal Identifier
scopus
DOI Link
https://doi.org/10.1109/ICME.2016.7552884
Copyright Status
Unknown
Socpus ID
84987643808 (Scopus)
Source API URL
https://api.elsevier.com/content/abstract/scopus_id/84987643808
STARS Citation
Li, Kai; Qi, Guojun; Ye, Jun; and Hua, Kien A., "Cross-Modal Hashing Through Ranking Subspace Learning" (2016). Scopus Export 2015-2019. 4003.
https://stars.library.ucf.edu/scopus2015/4003