Abstract
The research of hand pose estimation is a hot topic in computer vision, robotics and virtual reality. Compared with using data glove, vision based methods show great advantage for its contactless property, low-cost and convenience. With the commercial depth cameras became widely available and the great success of Convolution Neural Network (CNN) on images, various works focused on hand pose estimation have achieved promising performance. This research is inspired by the recent work that directly perform 3D classification and segmentation tasks on point cloud. In this paper, Multi-View PointNet (MVPointNet) is proposed which takes several views of point cloud as input source. Then, they fed into the well-performing point cloud-based architecture. In addition, to better capture the hand context structure and improve the performance, more features between centroid and local neighborhood points (norm, edge, angle) are extracted and fed into a deep CNN architecture. To our knowledge, our proposed method achieved good performance on the ModelNet40 dataset for 3D shape classification. Besides, it achieved superior performance over other deep learning methods for 3D hand pose estimation based on point cloud, which is evaluated on MSRA dataset.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.