Abstract
In recent years, hashing based cross-modal retrieval methods have attracted considerable attention, due to the significant reduction in computational cost and storage consumption. Most previous cross-modal hashing methods usually assume that data examples in different modalities are fully-paired. However, they neglect the fact that the data are often unpaired without one-to-one corresponding relationships in practical applications. Though several methods have noted the semi-paired or partial paired scenario, they ignore the completely unpaired scenario. In this paper, we propose a novel cross-modal hashing method, named Unpaired Cross-Modal Hashing (UCMH) for cross-modal retrieval to address the data with completely unpaired relationships. It leverages matrix factorization, similarity preservation, and semantic information to map data of different modalities to their respective semantic spaces. Moreover, different from most previous approaches, we construct an affinity matrix to bridge the semantic gap of data in different semantic spaces, which allows our method to handle single-label and multi-label unpaired cases simultaneously. Extensive experiments on one single-label dataset Wiki and two multi-label datasets namely MIR Flickr and NUS-WIDE, demonstrate that UCMH outperforms the state-of-the-art cross-modal hashing methods.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.