Abstract

This paper introduces two deep convolutional neural network training techniques that lead to more robust feature subspace separation in comparison to traditional training. Assume that dataset has M labels. The first method creates M deep convolutional neural networks called {text {DCNN}_i}_{i=1}^{M}. Each of the networks text {DCNN}_i is composed of a convolutional neural network (text {CNN}_i) and a fully connected neural network (text {FCNN}_i). In training, a set of projection matrices {mathbf {P}_i}_{i=1}^M are created and adaptively updated as representations for feature subspaces {mathcal {S}_i}_{i=1}^M. A rejection value is computed for each training based on its projections on feature subspaces. Each text {FCNN}_i acts as a binary classifier with a cost function whose main parameter is rejection values. A threshold value t_i is determined for i^{th} network text {DCNN}_i. A testing strategy utilizing {t_i}_{i=1}^M is also introduced. The second method creates a single DCNN and it computes a cost function whose parameters depend on subspace separations using the geodesic distance on the Grasmannian manifold of subspaces mathcal {S}_i and the sum of all remaining subspaces {mathcal {S}_j}_{j=1,jne i}^M. The proposed methods are tested using multiple network topologies. It is shown that while the first method works better for smaller networks, the second method performs better for complex architectures.

Highlights

  • There is an explosion of deep learning applications since the reintroduction of Convolutional Neural Networks (CNNs) in image classification [1] in 2012 and successive years to ImageNet dataset [2,3,4]

  • The proposed method is tested with five different topologies and compared to the results obtained with traditional Deep Convolutional Neural Network (DCNN) approach

  • Three topologies are to assess the effect of size on the performance, one topology topology adds dropout layers to the fully connected network and one topology has only CNN but not Fully Connected Neural network (FCNN)

Read more

Summary

Introduction

There is an explosion of deep learning applications since the reintroduction of Convolutional Neural Networks (CNNs) in image classification [1] in 2012 and successive years to ImageNet dataset [2,3,4]. It has been successfully applied in self-driving cars for traffic related objects and person detection and classification [5], face recognition for social media platforms [6], natural language processing [7], and symbolic mathematics [8]. DeepMind recently announced that their algorithm called AlphaFold can predict protein structures with atomic accuracy using deep learning [11, 12]

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call