Abstract

Abstract. With the rapid development of new indoor sensors and acquisition techniques, the amount of indoor three dimensional (3D) point cloud models was significantly increased. However, these massive “blind” point clouds are difficult to satisfy the demand of many location-based indoor applications and GIS analysis. The robust semantic segmentation of 3D point clouds remains a challenge. In this paper, a segmentation with layout estimation network (SLENet)-based 2D–3D semantic transfer method is proposed for robust segmentation of image-based indoor 3D point clouds. Firstly, a SLENet is devised to simultaneously achieve the semantic labels and indoor spatial layout estimation from 2D images. A pixel labeling pool is then constructed to incorporate the visual graphical model to realize the efficient 2D–3D semantic transfer for 3D point clouds, which avoids the time-consuming pixel-wise label transfer and the reprojection error. Finally, a 3D-contextual refinement, which explores the extra-image consistency with 3D constraints is developed to suppress the labeling contradiction caused by multi-superpixel aggregation. The experiments were conducted on an open dataset (NYUDv2 indoor dataset) and a local dataset. In comparison with the state-of-the-art methods in terms of 2D semantic segmentation, SLENet can both learn discriminative enough features for inter-class segmentation while preserving clear boundaries for intra-class segmentation. Based on the excellence of SLENet, the final 3D semantic segmentation tested on the point cloud created from the local image dataset can reach a total accuracy of 89.97%, with the object semantics and indoor structural information both expressed.

Highlights

  • In recent years, the location-based services (LBS) and GIS applications have been extended from outdoor to indoor environments (Zhou et al, 2017), which induces the rapid development of indoor data acquisition methodologies and sensors

  • The indoor 3D scenes created by light detection and ranging (LiDAR) surveying, consumer-level RGB-D camera collection, or structure from motion (SFM) technique are widely available (Dimitrov, Mani, 2015)

  • This paper proposes a new network, called segmentation with layout estimation network (SLENet) (Segmentation with Layout Estimation Network), to combine indoor semantic segmentation with space layout estimation

Read more

Summary

Introduction

The location-based services (LBS) and GIS applications have been extended from outdoor to indoor environments (Zhou et al, 2017), which induces the rapid development of indoor data acquisition methodologies and sensors. The indoor 3D scenes created by light detection and ranging (LiDAR) surveying, consumer-level RGB-D camera collection, or structure from motion (SFM) technique are widely available (Dimitrov , Mani, 2015). These 3D scenes are usually represented as semantically unkown 3D point clouds (Hermans et al, 2014). With the development of deep learning in 2D image classification, these methods have been progressively extended to 3D point cloud classification, becoming an important approach for 3D point cloud semantic

Objectives
Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call