Abstract

Deep learning models combining spectral and spatial features have been proven to be effective for hyperspectral image (HSI) classification. However, most spatial feature integration methods only consider a single input spatial scale regardless of various shapes and sizes of objects over the image plane, leading to missing scale-dependent information. In this paper, we propose a hierarchical multi-scale convolutional neural networks (CNNs) with auxiliary classifiers (HMCNN-AC) to learn hierarchical multi-scale spectral–spatial features for HSI classification. First, to better exploit the spatial information, multi-scale image patches for each pixel are generated at different spatial scales. These multi-scale patches are all centered at the same central spectrum but with shrunken spatial scales. Then, we apply multi-scale CNNs to extract spectral–spatial features from each scale patch. The obtained multi-scale convolutional features are considered as structured sequential data with spectral–spatial dependency, and a bidirectional LSTM is proposed to capture the correlation and extract a hierarchical representation for each pixel. To better train the whole network, weighted auxiliary classifiers are employed for the multi-scale CNNs and optimized together with the main loss function. Experimental results on three public HSI datasets demonstrate the superiority of our proposed framework over some state-of-the-art methods.

Highlights

  • Hyperspectral images (HSIs) [1] often contain hundreds of spectral bands varying from visible wavelength to short infrared light

  • To demonstrate the effectiveness of our proposed HMCNN-AC, several classification approaches are adopted. These compared approaches are based on two kinds of groups: one is based on classical machine learning methods, which includes (RBF-)support vector machine (SVM) and extended morphological profiles (EMP) [15]; the other is based on deep learning models, which includes stacked autoencoders (SAEs)-principal component analysis (PCA) [20], convolutional neural networks (CNNs)-multinomial logistic regression (MLR) [22], LSTM [25], and multi-scale convolutional neural networks (MCNNs) [28]

  • Unlike the conventional spectral–spatial classification methods where only one single input scale is considered for spatial feature integration, our proposed method could extract spectral–spatial features at various input scales simultaneously by adopting the multi-scale

Read more

Summary

Introduction

Hyperspectral images (HSIs) [1] often contain hundreds of spectral bands varying from visible wavelength to short infrared light This rich information enables us to distinguish different materials which look similar to the naked eye or the conventional RGB cameras. Pooling layers are often after the convolutional layers to further reduce the redundancy by partitioning the input data into a set of non-overlapped sub-regions and returning the average or maximum values locally. Both convolutional layers and pooling layers can be repeated multiple times to obtain the representative features. A fully connected layer is followed to further process the extracted features and a multinomial logistic regression (MLR) layer is added at the end to convert the output convolutional features into category or regression results

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call