Abstract

Land use and land cover (LULC) classification plays a significant role in Earth observation tasks. Nowadays, we can observe the same scene with multiple heterogeneous sensors. Combining diverse information therein for multisource joint classification has become a promising research topic in the remote sensing community. For example, the fusion of hyperspectral image (HSI) and lidar detection and ranging (LiDAR) data has been under active research. The current methodology for HSI and LiDAR joint classification tends to ignore the topological relationship between pixels, limiting the effectiveness of feature extraction and fusion. Another obstacle to satisfactory performance is the scarcity of annotated data. To overcome the above challenges, this article proposes a multisource attention network called MAGE to improve the collective classification. We use a semi-supervised graph transductive module to underline the relevance among pixels by explicitly constructing a multimodal adjacency matrix. Specifically, MAGE designs a self-supervised feature extraction module for pre-training, mitigating the dependence on annotated samples and alleviating the common overfitting and over-smoothing problems encountered by the deep graph neural network (GNN). The experimental results of three standard datasets, i.e., MUUFL, Trento, and Houston, demonstrate the effectiveness of the proposed approach. In particular, MAGE achieves an overall accuracy of 95.26% and an average accuracy of 96.27% on the challenging MUUFL dataset, surpassing the state-of-the-art methods. The code and models are publicly available at https://github.com/d1x1u/MAGE.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.