Abstract
Semantic segmentation of large-scale point clouds in 3D computer vision is a challenging problem. Existing feature extraction modules often emphasize learning local geometry while not giving adequate consideration to the integration of color information. This limitation prevents the network from thoroughly learning local features, thereby impacting segmentation accuracy. In this study, we propose three modules for robust feature extraction and aggregation, forming a novel point cloud segmentation network (HFA-Net) for large-scale point cloud semantic segmentation. First, we introduce the Hybrid Feature Extraction Component (HFEC) and the Hybrid Bilateral Enhancement Component (HBAC) to comprehensively extract and enhance the geometric, color, and semantic information of point clouds. Second, we incorporate the Ternary-Distance Attention Pooling (TDAP) module, which leverages trilateral distances to further refine the network’s focus on various features, enabling it to emphasize both locally important features and broader local neighborhoods. These modules are stacked into dense residual components to expand the network’s receptive field. Our experiments on several large-scale benchmark datasets, including Semantic3D, Toronto3D, S3DIS and LASDU demonstrate the effectiveness of HFA-Net when compared to state-of-the-art networks.
Published Version
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have