Abstract

Aiming at the problems of holes, misclassification, and rough edge segmentation in building extraction results from high spatial remote sensing images, a coordinate attention mechanism fusion network based on the BASNet network (CA-BASNet) is designed for building extraction in high spatial remote sensing images. Firstly, the deeply supervised encoder–decoder network was used to create a rough extract of buildings; secondly, to make the network pay more attention to learning building edge features, the mixed loss function composed of binary cross entropy, structural similarity and intersection-over-union was introduced into the network training process; finally, the residual optimization module of fusion coordinate attention mechanism was used for post-processing to realize the fine extraction of buildings from high spatial resolution remote sensing images. Experiments on the WHU building dataset show that the proposed network can achieve mIoU of 93.43%, mPA of 95.86%, recall of 98.79%, precision of 90.13% and F1 of 91.35%. Compared with the existing semantic segmentation networks, such as PSPNet, SegNet, DeepLapV3, SE-UNet, and UNet++, the accuracy of the proposed network and the integrity of object edge segmentation are significantly improved, which proves the effectiveness of the proposed network.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call