Abstract
Person tracking is an important issue in both computer vision and robotics. However, most existing person tracking methods using 3D point cloud are based on the Bayesian Filtering framework which are not robust in challenging scenes. In contrast with the filtering methods, in this paper, we propose a neural network to cope with person tracking using only 3D point cloud, named Point Siamese Network (PSN). PSN consists of two input branches named template and search, respectively. After finding the target person (by reading the label or using a detector), we get the inputs of the two branches and create feature spaces for them using feature extraction network. Meanwhile, a similarity map based on the feature space is proposed between them. We can obtain the target person from the map. Furthermore, we add an attention module to the template branch to guide feature extraction. To evaluate the performance of the proposed method, we compare it with the Unscented Kalman Filter (UKF) on 3 custom labeled challenging scenes and the KITTI dataset. The experimental results show that the proposed method performs better than UKF in robustness and accuracy and has a real-time speed. In addition, we publicly release our collected dataset and the labeled sequences to the research community.
Highlights
Person tracking is a key issue in both computer vision and mobile robotics
Inspired by the 2D image tracking method [2], we propose a neural network for point cloud tracking based on similarity computation, named Point Siamese
We propose an end-to-end network to handle person tracking only using point cloud, which has higher robustness than Unscented Kalman Filter (UKF)
Summary
Person tracking is a key issue in both computer vision and mobile robotics. For person following robots, they usually need to know the position of target person to follow. Person tracking is necessary for these robots. Most person tracking approaches are based on visual information [1,2,3]. Cameras have many advantages for tracking problems, such as they are compact and cheap, and they can provide abundant information. Visual tracking has its own limitations in practice, especially for autonomous driving car and target following robot
Published Version (Free)
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have