Abstract
In this paper, we concentrate on how to preserve fine-grained geometric structure information when extracting local contextual features for efficient large-scale point clouds semantic segmentation. Firstly, the Local Geometric Structure Representation Block is proposed to model fine-grained geometric structures for individual points by fully utilizing relative and global geometric relationships in the neighborhood. Then, we design a Parallel Attentive Fusion Module focusing on geometric structure and semantic information respectively, which reduces the feature’s ambiguity and preserves local geometric structure information. Furthermore, thanks to these two modules, we present a lightweight Local Contextual Features Extractor through the bilateral structure to obtain more discriminate local contextual features. Finally, a deep network named LGS-Net is introduced to predict point’s classes. Extensive experiment shows that our network surpasses the state-of-the-art approaches for semantic segmentation on two public large-scale point clouds datasets Semantic3D, SensatUrban, and achieve competitive performance on S3DIS. Especially, our LGS-Net with minimal model size outperforms the state-of-the-art network by 1.2% on the Semantic3D dataset. Thorough ablation studies and visualizations are presented to understand our network.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.