Abstract
Feature fusion methods have been demonstrated to be effective for many computer vision based applications. These methods generally use multiple hand-crafted features. However, in recent days, features extracted through transfer leaning procedures have been proved to be robust than the hand-crafted features in myriad applications, such as object classification and recognition. The transfer learning is a highly appreciated strategy in deep convolutional neural networks (DCNNs) due to its multifaceted benefits. It heartens us to explore the effect of fusing multiple transfer learning features of different DCNN architectures. Thus, in this work, we extract features of image statistics by exploiting three different pre-trained DCNNs through transfer learning. Then, we transform the features into a generalized subspace through a recently introduced Autoencoder network and fuse them to form intra-class invariant feature vector that is used to train a multi-class Support Vector Machine (SVM). The experimental results on various datasets, including object and action image statistics show that the fusion of multiple transfer learning features improves classification accuracy as compared to fusion of multiple hand-crafted features and usage of single component transfer learning features.
Published Version
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have