Abstract

With the ever increasing availability of various kinds of multimedia data, cross-modal retrieval, which enables information retrieval from various types of data given various types of query, has become a research hotspot. Hashing-based techniques have been developed to solve this problem, however, most previous works cannot capture the shared underlying structure of real-world multimodal data, which degrades their retrieval performances. In this paper, we propose a novel hashing method based on the extraction of the common manifold structure shared among different feature spaces. To faithfully represent the common structure, two kinds of local topology information are exploited in our method. Local angles are incorporated within the extraction of local topology of each feature space, which is then used to learn a common intermediate subspace. After heterogeneous features being embedded into this subspace, local similarities are exploited to extract the local topology between different feature spaces, and learn compact Hamming embeddings to facilitate cross-modal retrieval. The proposed method is referred to as full-space local topology extraction for hashing. Extensive comparisons with other state-of-the-art methods on three benchmark multimedia data sets demonstrate the superiority of our proposed method in terms of retrieval recall and search accuracy.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.