Abstract
In this paper a novel approach for multimodal search and retrieval is introduced. The searchable items are media representations consisting of multiple modalities, such as 2D images and 3D objects, which share a common semantic concept. The proposed method combines the low-level feature distances of each separate modality to construct a new low-dimensional feature space, where all media objects are mapped irrespective of their constituting modalities. While most of the existing state-of-the-art approaches support queries of one single modality at a time, the proposed one allows querying with multiple modalities simultaneously, through efficient multimodal query formulation.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.