Abstract

Abstract. Deep learning-based semantic segmentation models for building delineation face the challenge of producing precise and regular building outlines. Recently, a building delineation method based on frame field learning was proposed by Girard et al. (2020) to extract regular building footprints as vector polygons directly from aerial RGB images. A fully convolution network (FCN) is trained to learn simultaneously the building mask, contours, and frame field followed by a polygonization method. With the direction information of the building contours stored in the frame field, the polygonization algorithm produces regular outlines accurately detecting edges and corners. This paper investigated the contribution of elevation data from the normalized digital surface model (nDSM) to extract accurate and regular building polygons. The 3D information provided by the nDSM overcomes the aerial images’ limitations and contributes to distinguishing the buildings from the background more accurately. Experiments conducted in Enschede, the Netherlands, demonstrate that the nDSM improves building outlines’ accuracy, resulting in better-aligned building polygons and prevents false positives. The investigated deep learning approach (fusing RGB + nDSM) results in a mean intersection over union (IOU) of 0.70 in the urban area. The baseline method (using RGB only) results in an IOU of 0.58 in the same area. A qualitative analysis of the results shows that the investigated model predicts more precise and regular polygons for large and complex structures.

Highlights

  • Building extraction has been active for decades due to the availability of a large amount of very high-resolution remote sensing data and the need for detailed information of small-scale objects in multiple applications

  • AHN3 dataset is acquired in the 3rd acquisition period (2014-2019), and the DTM and Digital Surface Model (DSM) are derived from point cloud based on the Squared IDW method with 0.5 m resolution

  • We compared results obtained on the test set of aerial images (RGB) and composite images (RGB + normalized digital surface model (nDSM)) for the entire study area and for the urban area, respectively

Read more

Summary

Introduction

Building extraction has been active for decades due to the availability of a large amount of very high-resolution remote sensing data and the need for detailed information of small-scale objects in multiple applications. (iii) The high intra-class and low inter-class variation of building objects in high-resolution remotely sensed images make it hard to extract the buildings’ spectral and geometrical features (Huang, Zhang, Xin, Sun, & Zhang, 2019). To differentiate buildings from their complex background in Very High Resolution (VHR) remotely sensed images, a boundary refinement block (BRB) is introduced to amplify the distinction of features. Such a method’s performance decreases significantly with large buildings, resulting in less accurate outlines than Mask R-CNN (Li et al, 2019). The frame field and interior map are used in their polygonization algorithm to produce regular and accurate building polygons

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call