Cross-Modal Hashing Through Ranking Subspace Learning

Keywords

crossmodal hashing; Multimedia retrieval; ranking subspace learning; WTA hash

Abstract

Hashing has been widely used for approximate nearest neighbor search of high-dimensional multimedia data. In this paper, we propose a novel hash learning framework that maps high-dimensional multimodal data into a common Hamming space where the cross-modal similarity can be measured using Hamming distance. Unlike existing cross-modal hashing methods that learn hash functions in the form of numeric quantization of linear projections, the proposed hash learning algorithm encodes features' ranking properties and takes advantage of rank correlations which are known to be scale-invariant, numerically stable and highly nonlinear. Specifically, we learn two groups of subspaces jointly, one for each modality, so that the ranking orders in those subspaces maximally preserve the cross-modal similarity. Extensive experiments on realworld datasets demonstrate superiority of the proposed methods compared to state-of-the-arts.

Publication Date

8-25-2016

Publication Title

Proceedings - IEEE International Conference on Multimedia and Expo

Volume

2016-August

Document Type

Article; Proceedings Paper

Personal Identifier

scopus

DOI Link

https://doi.org/10.1109/ICME.2016.7552884

Socpus ID

84987643808 (Scopus)

Source API URL

https://api.elsevier.com/content/abstract/scopus_id/84987643808

This document is currently not available here.

Share

COinS