Abstract

Material images are susceptible to changes, depending on the light intensity, visual angle, shooting distance, and other conditions. Feature learning has shown great potential for addressing this issue. However, the knowledge achieved using a simple feature fusion method is insufficient to fully represent the material images. In this study, we aimed to exploit the diverse knowledge learned by a novel progressive feature fusion method to improve the recognition performance. To obtain implicit cross-modal knowledge, we perform early feature fusion and capture the cluster canonical correlations among the state-of-the-art (SOTA) heterogeneous squeeze-and-excitation network (SENet) features. A set of more discriminative deep-level visual semantics (DVSs) is obtained. We then perform gene selection-based middle feature fusion to thoroughly exploit the feature-shared knowledge among the generated DVSs. Finally, any type of general classifier can use the feature-shared knowledge to perform the final material recognition. Experimental results on two public datasets (Fabric and MattrSet) showed that our method outperformed other SOTA baseline methods in terms of accuracy and real-time efficiency. Even most traditional classifiers were able to obtain a satisfactory performance using our method, thus demonstrating its high practicality.

Highlights

  • Material recognition [1,2,3,4,5] has become a popular topic in the field of computer vision (CV)

  • Each sample includes four images; it has a total of 5064 images (Kampouris et al cropped all the images to 400 × 400 pixels to avoid blurring the edges of the images)

  • (2) We found that, regardless of the number of deep-level visual semantics (DVSs), the support vector machine (SVM) classifier always achieved the best performance on the MattrSet dataset and the logistic regression (LR) classifier on the Fabric dataset. ese findings are further illustrated in Figure 4. ey indicate that the most suitable classifier should be chosen carefully for different material datasets to achieve the largest performance improvement

Read more

Summary

Introduction

Material recognition [1,2,3,4,5] has become a popular topic in the field of computer vision (CV). It has significant value in many practical scenarios, such as scene recognition [6], industrial inspection [7], medical image recognition [8], and robot vision [9]. Cleaning robots must distinguish between wood, tiles, and carpets. A computer-aided material recognition model can address this issue well and applies to the aforementioned practical fields; it is still a great challenge, as material recognition is normally affected by several external factors, including light intensity, visual angle, shooting distance, and other conditions. Feature learning has shown great potential for alleviating this problem

Objectives
Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call