To address the challenges of limited detection precision and insufficient segmentation of small to medium-sized objects in dynamic and complex scenarios, such as the dense intermingling of pedestrians, vehicles, and various obstacles in urban environments, we propose an enhanced methodology. Firstly, we integrated a point cloud processing module utilizing the DBSCAN clustering algorithm to effectively segment and extract critical features from the point cloud data. Secondly, we introduced a fusion attention mechanism that significantly improves the network’s capability to capture both global and local features, thereby enhancing object detection performance in complex environments. Finally, we incorporated a CSPNet downsampling module, which substantially boosts the network’s overall performance and processing speed while reducing computational costs through advanced feature map segmentation and fusion techniques. The proposed method was evaluated using the KITTI dataset. Under moderate difficulty, the BEV mAP for detecting cars, pedestrians, and cyclists achieved 87.74%, 55.07%, and 67.78%, reflecting improvements of 1.64%, 5.84%, and 5.53% over PointPillars. For 3D mAP, the detection accuracy for cars, pedestrians, and cyclists reached 77.90%, 49.22%, and 62.10%, with improvements of 2.91%, 5.69%, and 3.03% compared to PointPillars.
Read full abstract