Abstract
This paper presents a novel method to determine the discriminative image representation via visual dictionary learning framework for image classification task. Visual dictionary learning has the capacity to represent input image using an over-complete element set. Sparsity restrains distractors and prevents over-fitting. The two main characteristics benefit the classification solution. However, one shortcoming of existing dictionary learning is that it neglects to exploit the potential correlations across visual elements, especially from the category-specific feature space. To address this problem, we first propose to learn multiple discriminative category-specific dictionaries (DCSD) from all categories. The DCSD can explore the visual elements from each category in terms of sharable property. For this reason, these learned category-specific visual elements encourage image features from the same class to have the similar feature representations. In addition, exemplary data reflect the main characteristic of whole dataset and can improve the performance of algorithm that employs them. Therefore, we further propose a representative pattern dictionary (RPD) model to discover the exemplary visual elements for promoting the discriminative capability of feature representation. These exemplary visual elements are essentially a subset of over-complete visual elements and can represent the whole sample data effectively. Finally, we design a novel strategy that integrates the merits of object proposals and deep features jointly to strengthen the semantic information of image-level feature. Experimental results on benchmark datasets demonstrate the effectiveness of our method, which is shown to be superior to the recently competing dictionary learning and deep learning based image classification approaches.
Highlights
Image classification is a fundamental problem in computer vision and has attracted significant attentions
Inspired by the main idea of dictionary learning, we propose to discover the category-specific sharable and exemplary visual elements by means of the dictionary learning framework, which shows superior performance compared with recently competing dictionary learning and deep learning approaches
The goal of our method aims to discover the category-specific sharable and exemplary visual elements by learning multiple discriminative category-specific dictionaries (DCSD) and a representative pattern dictionary (RPD) across all categories, respectively
Summary
Image classification is a fundamental problem in computer vision and has attracted significant attentions. In order to enrich the semantic information for feature representation, we first propose to learn a set of discriminative category-specific dictionaries (DCSD) in terms of sharable visual patterns. To further enhance the semantic information of feature representation, we propose a new representative pattern dictionary learning (RPD) model to explore the exemplary visual patterns across all categories. These exemplary visual patterns denote a subset of an over-complete visual elements and can reconstruct the whole visual data effectively.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.