Abstract
The acquisition of multi-view hash representation for heterogeneous data holds paramount importance in the domain of multimedia retrieval. The limited retrieval precision observed in current approaches stems from their inadequate integration of multi-view features and their failure to effectively leverage the metric information available from diverse samples. Commonly employed fusion methods, such as concatenation or weighted sum, are insufficient in capturing the complementarity among multiple view features. Furthermore, these methods neglect the valuable information contributed by dissimilar samples. To address these challenges, we propose an innovative method termed Fast Metric Multi-View Hashing (FMMVH). Our approach showcases the superiority of gate-based fusion over traditional methods, as substantiated by extensive empirical evidence. Additionally, this paper proposes a novel deep metric loss function to enable the utilization of metric information from dissimilar samples. We exclusively train our method using this single loss function. To enhance practical applicability in industrial production environments, we employ model compression techniques to optimize the proposed method. On benchmark datasets such as MIR-Flickr25K, NUS-WIDE, and MS COCO, the performance of our FMMVH method significantly surpasses that of existing state-of-the-art methods, demonstrating improvements of up to 7.47% in mean Average Precision (mAP).
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.