Abstract

Land cover classification of high-resolution remote sensing images aims to obtain pixel-level land cover understanding, which is often modeled as semantic segmentation of remote sensing images. In recent years, convolutional network (CNN)-based land cover classification methods have achieved great advancement. However, previous methods fail to generate fine segmentation results, especially for the object boundary pixels. In order to obtain boundary-preserving predictions, we first propose to incorporate spatially adapting contextual cues. In this way, objects with similar appearance can be effectively distinguished with the extracted global contextual cues, which are very helpful to identify pixels near object boundaries. On this basis, low-level spatial details and high-level semantic cues are effectively fused with the help of our proposed dual attention mechanism. Concretely, when fusing multi-level features, we utilize the dual attention feature fusion module based on both spatial and channel attention mechanisms to relieve the influence of the large gap, and further improve the segmentation accuracy of pixels near object boundaries. Extensive experiments were carried out on the ISPRS 2D Semantic Labeling Vaihingen data and GaoFen-2 data to demonstrate the effectiveness of our proposed method. Our method achieves better performance compared with other state-of-the-art methods.

Highlights

  • Accepted: 14 September 2021With the development of very high resolution (VHR) remote sensing technology, large amounts of satellite remote sensing images with very high resolution are obtained every day [1]

  • With the development of deep learning [10,11,12,13,14], there are lots of convolutional network (CNN)-based methods [15,16,17,18,19,20,21] applied in the semantic segmentation of the VHR remote sensing images

  • We first present a general framework of our network and introduce the adaptive context aggregating module (ACAM) and the dual attention feature fusion module (DAFFM)

Read more

Summary

Introduction

Accepted: 14 September 2021With the development of very high resolution (VHR) remote sensing technology, large amounts of satellite remote sensing images with very high resolution are obtained every day [1]. There are still lots of challenges for the task of semantic segmentation in VHR remote sensing images with complex scenes, such as poor accuracy of multi-category semantic segmentation, poor speed of multi-category semantic segmentation, and so on. With the development of deep learning [10,11,12,13,14], there are lots of CNN-based methods [15,16,17,18,19,20,21] applied in the semantic segmentation of the VHR remote sensing images. In order to show the effectiveness of the proposed method, we further perform comparisons with state-of-the-art semantic segmentation methods on both ISPRS 2D Semantic. Results based on ISPRS 2D Semantic Labeling Vaihingen data [35].

Objectives
Methods
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call