Abstract

3D object detection is a challenging task in autonomous driving industry scenarios. Many pre-existing methods employ the set-abstraction operation for generating key-point representations, which, however, cannot learn the long-range context dependency properly. In addition, the pooling operator, which only focuses on maximum channel response, is adopted to aggregate features of neighbor points without semantic information. To fix these issues, we propose LGNet, a new framework that simultaneously captures local and global point dependencies for enhancing 3D object detection. Specifically, we first introduce a new local point-graph pooling module to compute point-to-point correlations in a local region and aggregate features from neighboring points. To further capture the long-range dependency in a global context, we devised a global point-aware module to integrate local and global features at higher resolution. Experiments on the KITTI 3D detection dataset and Waymo Open Dataset benchmark show that LGNet achieves state-of-the-art performance in multiple classes. We will upload the code on https://github.com/MWPony/LGNet.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.