ABSTRACT High-resolution remote sensing (HRRS) scene classification is based on the semantic information in the image and its contextual spatial correlation, labelling different semantic categories for different HRRS. Convolutional neural networks (CNN) have been widely studied and applied in remote sensing scene classification in recent years. However, most of the existing CNN models focus on the global and high-level semantic features, ignoring the shallower feature information. In addition, complex background information and variable scales lead to a series of problems with large intra-class differences and high inter-class similarities, which also brings challenges to scene classification. To address the above problems and challenges, a scenario classification model based on the contextual spatial attention and channel attention mechanism of Lie Group manifold space learning is proposed in this study. In this model, we fully explore the multi-scale features of the scene (shallower-level and high-level) and propose a novel contextual spatial attention mechanism and channel attention mechanism. Extensive experimentation was carried out on the Union Remote Sensing Image Data Set (URSIS), which improved by 7.13% compared with the classical model. The experimental results showed that compared with other state-of-the-art remote sensing scene classification models, our proposed method has achieved significant improvement in classification accuracy and performance.
Read full abstract