Coarse-Refined Local Attention Network for Hyperspectral Image Classification

  • Abstract
  • Literature Map
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

Recently, deep learning methods using the attention mechanism have generated considerable research interest for hyperspectral image classification. In many existing attention-based methods, global pooling is widely used to obtaining the attention weights. In general, there are multiple categories in a hyperspectral image, so the operation of global pooling is too crude and inappropriate. To alleviate this problem, we propose a coarse-refined local attention network (CRLAN) for hyperspectral image classification. CRLAN is composed of two stages of fully convolutional networks. The first stage employs a coarse local attention fully convolutional network for hyperspectral image classification. In this stage, local parameters are roughly estimated according to the original size of the hyperspectral image. In the second stage, the prediction classification probability of the first stage network is applied to obtain the refined local attention features. Finally, for testing convenience, these two stages are integrated into an end-to-end network. Experimental results on two public data sets demonstrate that CRLAN is effective in improving classification performance.

Similar Papers
  • PDF Download Icon
  • Research Article
  • Cite Count Icon 32
  • 10.3390/rs12122035
Residual Group Channel and Space Attention Network for Hyperspectral Image Classification
  • Jun 24, 2020
  • Remote Sensing
  • Peida Wu + 3 more

Recently, deep learning methods based on three-dimensional (3-D) convolution have been widely used in the hyperspectral image (HSI) classification tasks and shown good classification performance. However, affected by the irregular distribution of various classes in HSI datasets, most previous 3-D convolutional neural network (CNN)-based models require more training samples to obtain better classification accuracies. In addition, as the network deepens, which leads to the spatial resolution of feature maps gradually decreasing, much useful information may be lost during the training process. Therefore, how to ensure efficient network training is key to the HSI classification tasks. To address the issue mentioned above, in this paper, we proposed a 3-DCNN-based residual group channel and space attention network (RGCSA) for HSI classification. Firstly, the proposed bottom-up top-down attention structure with the residual connection can improve network training efficiency by optimizing channel-wise and spatial-wise features throughout the whole training process. Secondly, the proposed residual group channel-wise attention module can reduce the possibility of losing useful information, and the novel spatial-wise attention module can extract context information to strengthen the spatial features. Furthermore, our proposed RGCSA network only needs few training samples to achieve higher classification accuracies than previous 3-D-CNN-based networks. The experimental results on three commonly used HSI datasets demonstrate the superiority of our proposed network based on the attention mechanism and the effectiveness of the proposed channel-wise and spatial-wise attention modules for HSI classification. The code and configurations are released at Github.com.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 8
  • 10.3390/rs13071290
Wide Sliding Window and Subsampling Network for Hyperspectral Image Classification
  • Mar 28, 2021
  • Remote Sensing
  • Jiangbo Xi + 6 more

Recently, deep learning methods, for example, convolutional neural networks (CNNs), have achieved high performance in hyperspectral image (HSI) classification. The limited training samples of HSI images make it hard to use deep learning methods with many layers and a large number of convolutional kernels as in large scale imagery tasks, and CNN-based methods usually need long training time. In this paper, we present a wide sliding window and subsampling network (WSWS Net) for HSI classification. It is based on layers of transform kernels with sliding windows and subsampling (WSWS). It can be extended in the wide direction to learn both spatial and spectral features more efficiently. The learned features are subsampled to reduce computational loads and to reduce memorization. Thus, layers of WSWS can learn higher level spatial and spectral features efficiently, and the proposed network can be trained easily by only computing linear weights with least squares. The experimental results show that the WSWS Net achieves excellent performance with different hyperspectral remotes sensing datasets compared with other shallow and deep learning methods. The effects of ratio of training samples, the sizes of image patches, and the visualization of features in WSWS layers are presented.

  • Research Article
  • Cite Count Icon 7
  • 10.1080/01431161.2021.1993464
Two-Stage Attention Network for hyperspectral image classification
  • Nov 6, 2021
  • International Journal of Remote Sensing
  • Peida Wu + 3 more

Considering that the hyperspectral image (HSI) has a large number of spectrum bands, to optimize the features and make full use of more informative features, many papers have introduced attention mechanism to the models based on three-dimensional (3D) convolution. However, though the number of spectrum bands is large, there are many useless bands and noise, which may generate lots of useless features into the subsequent network and affect the learning efficiency of each convolutional layer. Therefore, how to reduce the influence of noise from HSI data itself and the classification process is key to the HSI classification tasks. In this paper, we proposed a 3D convolutional neural network (3D-CNN) based two-stage attention network (TSAN) for HSI classification. For one thing, the spectral-wise attention module in the first stage can optimize the whole spectrum by shielding useless spectrum bands and reducing the noise in the spectrum. For another, more discriminative spectral–spatial features are extracted and sent to the subsequent layers by channel-wise attention mechanism combined with soft thresholding in the second stage. In addition, we introduced non-local block to learn global spatial features and used a multi-scale network to combine the local space and the global space. The experiments carried out on three HSI datasets show that our proposed network for HSI classification tasks can indeed reduce the noise by soft thresholding and achieve promising classification performance.

  • Research Article
  • 10.1080/01431161.2024.2398822
Cnn-assisted multi-hop graph attention network for hyperspectral image classification
  • Oct 3, 2024
  • International Journal of Remote Sensing
  • Hongxi Wang + 3 more

Recently, the convolutional neural network (CNN) has gained widespread adoption in the hyperspectral image (HSI) classification owing to its remarkable feature extraction capability. However, the fixed acceptance domain of CNN restricts it to Euclidean image data only, making it difficult to capture complex information in hyperspectral data. To overcome this problem, much attention has been paid to the graph attention network (GAT), which can effectively model graph structure and capture complex dependencies between nodes. However, GAT usually acts on superpixel nodes, which may lead to the loss of pixel-level information. To better integrate the advantages of both, we propose a CNN-assisted multi-hop graph attention network (CMGAT) for HSI classification. Specifically, a parallel dual-branch architecture is first constructed to simultaneously capture spectral-spatial features from hyperspectral data at the superpixel and pixel levels using GAT and CNN, respectively. On this basis, the multi-hop and multi-scale mechanisms are further employed to construct a multi-hop GAT module and a multi-scale CNN module to capture diverse feature information. Secondly, an attention module is cascaded before the multi-scale CNN module to improve classification performance. Eventually, the output information from the two branches is weighted and fused to produce the classification result. We performed experiments on four benchmark HSI datasets, including Indian Pines (IP), University of Pavia (UP), Salinas Valley (SV) and WHU-Hi-LongKou (LK). The results demonstrate that the proposed method outperforms several deep learning methods, achieving overall accuracies of 95.67%, 99.04%, 99.55% and 99.51%, respectively, even with fewer training samples.

  • Research Article
  • Cite Count Icon 189
  • 10.1080/2150704x.2017.1331053
A semi-supervised convolutional neural network for hyperspectral image classification
  • May 23, 2017
  • Remote Sensing Letters
  • Bing Liu + 5 more

ABSTRACTConvolutional neural network (CNN) for hyperspectral image classification can provide excellent performance when the number of labeled samples for training is sufficiently large. Unfortunately, a small number of labeled samples are available for training in hyperspectral images. In this letter, a novel semi-supervised convolutional neural network is proposed for the classification of hyperspectral image. The proposed network can automatically learn features from complex hyperspectral image data structures. Furthermore, skip connection parameters are added between the encoder layer and decoder layer in order to make the network suitable for semi-supervised learning. Semi-supervised method is adopted to solve the problem of limited labeled samples. Finally, the network is trained to simultaneously minimize the sum of supervised and unsupervised cost functions. The proposed network is conducted on a widely used hyperspectral image data. The experimental results demonstrate that the proposed approach provides competitive results to state-of-the-art methods.

  • Research Article
  • Cite Count Icon 9
  • 10.1080/2150704x.2019.1686780
A hybrid neural network for hyperspectral image classification
  • Nov 12, 2019
  • Remote Sensing Letters
  • Jiangyun Li + 2 more

ABSTRACTRecent research shows that deep learning-based methods can achieve promissing performance when applied to hyperspectral image (HSI) classification in remote sensing, some challenging issues still exist. For example, after a number of 2D convolutions, each feature map may only correspond to a unique dimension of the hyperspectral image. As a result, the relationship between different feature maps from multiple dimensional hyperspectral image can not be extracted well. Another issue is information in extracted feature maps may be erased by pooling operations. To address these problems, we propose a novel hybrid neural network (HNN) for hyperspectral image classification. The HNN uses a multi-branch architecture to extract hyperspectral image features in order to improve its prediction accuracy. Moreover, we build a deconvolution structure to recover the lost information in the pooling operation. In addition, to improve convergence and prevent overfitting, the HNN applies batch normalization (BN) and parametric rectified linear units (PReLU). In the experiments, two public benchmark HSIs are utilized to evaluate the performance of the proposed method. The experimental results demonstrate the superiority of HNN over several well-known methods.

  • Research Article
  • Cite Count Icon 13
  • 10.1080/01431161.2023.2249598
CNN and Transformer interaction network for hyperspectral image classification
  • Sep 8, 2023
  • International Journal of Remote Sensing
  • Zhongwei Li + 4 more

Convolutional Neural Network (CNN) has developed hyperspectral image (HSI) classification effectively. Although many CNN-based models can extract local features in HSI, it is difficult for them to extract global features. With its ability to capture long-range dependencies, Transformer is gradually gaining prominence in HSI classification, but it may overlook some local details when extracting features. To address these issues, we proposed a CNN and transformer interaction network (CTIN) for HSI classification. Firstly, A dual-branch structure was constructed in which CNN and Transformer are arranged in parallel to simultaneously extract global features and local features in HSI. Secondly, a feature interaction module has been imported between the two branches, thus facilitating a bi-directional flow of information between the global and local feature spaces. In this way, the network structure combines the advantages of CNN and Transformer in extracting features as much as possible. In addition, a token generation method is designed to harness abundant contextual information that is relevant to the centre pixel, and improve the accuracy of the final classification. Experiments were conducted on four hyperspectral datasets (two classical datasets – Indian Pines, Salinas Valley, a new satellite dataset – Yellow River, and an self-made UAV dataset-Yellow River Willow). Experimental results show that the proposed method outperforms the other state-of-the-art methods, with overall accuracies of 99.21%, 99.61%, 92.40%, and 98.17%, respectively.

  • Research Article
  • Cite Count Icon 2
  • 10.3390/rs16224202
SSFAN: A Compact and Efficient Spectral-Spatial Feature Extraction and Attention-Based Neural Network for Hyperspectral Image Classification
  • Nov 11, 2024
  • Remote Sensing
  • Chunyang Wang + 6 more

Hyperspectral image (HSI) classification is a crucial technique that assigns each pixel in an image to a specific land cover category by leveraging both spectral and spatial information. In recent years, HSI classification methods based on convolutional neural networks (CNNs) and Transformers have significantly improved performance due to their strong feature extraction capabilities. However, these improvements often come with increased model complexity, leading to higher computational costs. To address this, we propose a compact and efficient spectral-spatial feature extraction and attention-based neural network (SSFAN) for HSI classification. The SSFAN model consists of three core modules: the Parallel Spectral-Spatial Feature Extraction Block (PSSB), the Scan Block, and the Squeeze-and-Excitation MLP Block (SEMB). After preprocessing the HSI data, it is fed into the PSSB module, which contains two parallel streams, each comprising a 3D convolutional layer and a 2D convolutional layer. The 3D convolutional layer extracts spectral and spatial features from the input hyperspectral data, while the 2D convolutional layer further enhances the spatial feature representation. Next, the Scan Block module employs a layered scanning strategy to extract spatial information at different scales from the central pixel outward, enabling the model to capture both local and global spatial relationships. The SEMB module combines the Spectral-Spatial Recurrent Block (SSRB) and the MLP Block. The SSRB, with its adaptive weight assignment mechanism in the SToken Module, flexibly handles time steps and feature dimensions, performing deep spectral and spatial feature extraction through multiple state updates. Finally, the MLP Block processes the input features through a series of linear transformations, GELU activation functions, and Dropout layers, capturing complex patterns and relationships within the data, and concludes with an argmax layer for classification. Experimental results show that the proposed SSFAN model delivers superior classification performance, outperforming the second-best method by 1.72%, 5.19%, and 1.94% in OA, AA, and Kappa coefficient, respectively, on the Indian Pines dataset. Additionally, it requires less training and testing time compared to other state-of-the-art deep learning methods.

  • Research Article
  • Cite Count Icon 28
  • 10.1016/j.engappai.2023.107280
Fuzzy graph convolutional network for hyperspectral image classification
  • Oct 12, 2023
  • Engineering Applications of Artificial Intelligence
  • Jindong Xu + 6 more

Fuzzy graph convolutional network for hyperspectral image classification

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 43
  • 10.3390/rs12010125
A Multi-Scale and Multi-Level Spectral-Spatial Feature Fusion Network for Hyperspectral Image Classification
  • Jan 1, 2020
  • Remote Sensing
  • Caihong Mu + 2 more

Extracting spatial and spectral features through deep neural networks has become an effective means of classification of hyperspectral images. However, most networks rarely consider the extraction of multi-scale spatial features and cannot fully integrate spatial and spectral features. In order to solve these problems, this paper proposes a multi-scale and multi-level spectral-spatial feature fusion network (MSSN) for hyperspectral image classification. The network uses the original 3D cube as input data and does not need to use feature engineering. In the MSSN, using different scale neighborhood blocks as the input of the network, the spectral-spatial features of different scales can be effectively extracted. The proposed 3D–2D alternating residual block combines the spectral features extracted by the three-dimensional convolutional neural network (3D-CNN) with the spatial features extracted by the two-dimensional convolutional neural network (2D-CNN). It not only achieves the fusion of spectral features and spatial features but also achieves the fusion of high-level features and low-level features. Experimental results on four hyperspectral datasets show that this method is superior to several state-of-the-art classification methods for hyperspectral images.

  • Research Article
  • Cite Count Icon 13
  • 10.1016/j.displa.2021.102114
Discriminative graph convolution networks for hyperspectral image classification
  • Nov 10, 2021
  • Displays
  • Zhe Wang + 2 more

Discriminative graph convolution networks for hyperspectral image classification

  • Research Article
  • Cite Count Icon 14
  • 10.1080/2150704x.2021.1992034
A probabilistic neighbourhood pooling-based attention network for hyperspectral image classification
  • Nov 2, 2021
  • Remote Sensing Letters
  • Yuanlin Wang + 3 more

Attention mechanisms are recently deployed in deep learning models for hyperspectral image (HSI) classification. Conventional spectral attentions typically use global pooling to aggregate spatial information, without sufficiently considering the spatial dependencies of the central pixel to be classified and its neighbours. Moreover, the limited training samples with high-dimensional spectral information make deep learning models prone to over-fitting. In view of these, we propose an end-to-end probabilistic neighbourhood pooling-based attention network (PNPAN) for HSI classification. In PNPAN, we divided the feature maps of input HSI cubes into ring-shaped neighbouring regions and probabilistically selected them as pooling regions to compute channel-wise attention. Based on this, we built a spectral attention-based module and a 3-D convolution module to extract spectral-spatial features. Experiments on three benchmark data sets demonstrate that PNPAN achieves promising results for HSI classification with limited training samples.

  • Research Article
  • Cite Count Icon 3
  • 10.1364/josaa.478585
Hybrid spatial-spectral generative adversarial network for hyperspectral image classification.
  • Feb 21, 2023
  • Journal of the Optical Society of America A
  • Chao Ma + 5 more

In recent years, generative adversarial networks (GNAs), consisting of two competing 2D convolutional neural networks (CNNs) that are used as a generator and a discriminator, have shown their promising capabilities in hyperspectral image (HSI) classification tasks. Essentially, the performance of HSI classification lies in the feature extraction ability of both spectral and spatial information. The 3D CNN has excellent advantages in simultaneously mining the above two types of features but has rarely been used due to its high computational complexity. This paper proposes a hybrid spatial-spectral generative adversarial network (HSSGAN) for effective HSI classification. The hybrid CNN structure is developed for the construction of the generator and the discriminator. For the discriminator, the 3D CNN is utilized to extract the multi-band spatial-spectral feature, and then we use the 2D CNN to further represent the spatial information. To reduce the accuracy loss caused by information redundancy, a channel and spatial attention mechanism (CSAM) is specially designed. To be specific, a channel attention mechanism is exploited to enhance the discriminative spectral features. Furthermore, the spatial self-attention mechanism is developed to learn the long-term spatial similarity, which can effectively suppress invalid spatial features. Both quantitative and qualitative experiments implemented on four widely used hyperspectral datasets show that the proposed HSSGAN has a satisfactory classification effect compared to conventional methods, especially with few training samples.

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 50
  • 10.3390/rs12122033
Synergistic 2D/3D Convolutional Neural Network for Hyperspectral Image Classification
  • Jun 24, 2020
  • Remote Sensing
  • Xiaofei Yang + 6 more

Accurate hyperspectral image classification has been an important yet challenging task for years. With the recent success of deep learning in various tasks, 2-dimensional (2D)/3-dimensional (3D) convolutional neural networks (CNNs) have been exploited to capture spectral or spatial information in hyperspectral images. On the other hand, few approaches make use of both spectral and spatial information simultaneously, which is critical to accurate hyperspectral image classification. This paper presents a novel Synergistic Convolutional Neural Network (SyCNN) for accurate hyperspectral image classification. The SyCNN consists of a hybrid module that combines 2D and 3D CNNs in feature learning and a data interaction module that fuses spectral and spatial hyperspectral information. Additionally, it introduces a 3D attention mechanism before the fully-connected layer which helps filter out interfering features and information effectively. Extensive experiments over three public benchmarking datasets show that our proposed SyCNNs clearly outperform state-of-the-art techniques that use 2D/3D CNNs.

  • Research Article
  • Cite Count Icon 9
  • 10.1080/01431161.2021.1934599
Composite kernel learning network for hyperspectral image classification
  • Jun 16, 2021
  • International Journal of Remote Sensing
  • Zhe Wu + 4 more

The small sample problem has always been a serious challenge in hyperspectral image (HSI) classification. In order to obtain satisfactory results when the training samples are insufficient, the information around the training samples should be fully utilized. In this paper, we focus on small sample learning and propose a novel composite kernel learning network (CKLNet) for HSI classification. First, principal component analysis and extended morphological analysis are utilized to extract features. Then, we introduce generalized kernel method into deep learning technology. The spatial-spectral composite kernel learning (SSCKL) module is developed to construct discriminative and robust spatial-spectral generalized kernel features. In the process of constructing kernel features, the deep correlation information between samples is extracted simultaneously. The kernel hyperparameters in SSCKL are automatically learnt through backpropagation, thus avoiding the need to spend a lot of time on cross-validation. Finally, inspired by U-Net, a global-local feature extraction (GLFE) module is designed to extract spatial features of different scales. A set of classification probability maps can be obtained by the 11 convolutional layer in the GLFE module. Experimental results on three widely used datasets demonstrate the effectiveness of the proposed CKLNet.

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.