Abstract

Object recognition based on LIDAR data is crucial in automotive driving and is the subject of extensive research. However, the lack of accuracy and stability in complex environments obstructs the practical application of real-time recognition algorithms. In this study, we proposed a new real-time network for multicategory object recognition. The manually extracted bird’s eye view (BEV) features were adopted to replace the resource-consuming 3D convolutional operation. Besides the subject network, we designed two auxiliary networks to help the network learn the pointwise features and boxwise features, aiming to improve the category and bounding boxes’ accuracy. The KITTI dataset was adopted to train and validate the proposed network. Experimental results showed that, for hard mode, the total average precision (AP) of the category reached 97.4%. For an intersection over a union threshold of 0.5 and 0.7, the total AP of regression reached 93.2% and 85.5%; especially, the AP of car’s regression reached 95.7% and 92.2%. The proposed network also showed consistent performance in the Apollo dataset with a processing duration of 37 ms. The proposed network exhibits stable and robust object recognition performance in complex environments (multiobject, unordered objects, and multicategory). And it shows sensitivity to occlusion of the LIDAR system and insensitivity to close large objects. The proposed multifunction method simultaneously achieves real-time operation, high accuracy, and stable performance, indicating its great potential value in practical application.

Highlights

  • Autonomous driving is a futuristic technology that will transform mobility industries and ease the burden of driving

  • We evaluated the method based on the 3D localization and bounding box precision, object recognition accuracy, and processing time

  • Unlike most present methods that heavily rely on 3D convolutional operations, we considered that the bird’s eye view (BEV) based method has not yet exhausted its performance potential. us, we improved the head network and designed an additional auxiliary network to improve the prediction accuracy. e network was trained and evaluated by the KITTI dataset and its benchmark. e results verify that the new part is beneficial to the network

Read more

Summary

Research Article

Combined Auxiliary Networks and Bird’s Eye View Method for Real-Time Multicategory Object Recognition. Object recognition based on LIDAR data is crucial in automotive driving and is the subject of extensive research. The lack of accuracy and stability in complex environments obstructs the practical application of real-time recognition algorithms. We proposed a new real-time network for multicategory object recognition. Experimental results showed that, for hard mode, the total average precision (AP) of the category reached 97.4%. E proposed network showed consistent performance in the Apollo dataset with a processing duration of 37 ms. E proposed network exhibits stable and robust object recognition performance in complex environments (multiobject, unordered objects, and multicategory). E proposed multifunction method simultaneously achieves real-time operation, high accuracy, and stable performance, indicating its great potential value in practical application It shows sensitivity to occlusion of the LIDAR system and insensitivity to close large objects. e proposed multifunction method simultaneously achieves real-time operation, high accuracy, and stable performance, indicating its great potential value in practical application

Introduction
Dense Dense
Cp gt
Results and Discussion
Cyclist Total
Mode Car Truck Pedestrian Cyclist Total
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call