Abstract
In this paper, we propose a novel method for scene image retrieval in which the semantic meaning of an image and a new low-level feature are combined. The fluid nature of scene images makes learning semantics essential in our retrieval task. Compared to a general image, a scene image contains large regions of low contrast, which makes it difficult for a method to extract features that has good coverage of the entire image and assurance of relatively high repeatability. Given a scene image as a query, a collection of images is first retrieved by some search engines based on the images' semantic meanings. The candidate images are re-ranked by adapting an asymmetric piece-to-image matching scheme based on their visual similarities with the query image, using its visual signature consists of some packed dense interest points. Our method is evaluated on an Outdoor Scene Recognition (OSR) dataset and an NUS-WIDE dataset. It has demonstrated the improvements of our method over other conventional approaches.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.