Abstract
Accurate keypoint localization is an important research direction in the field of robotic arms. Deep learning is being used more and more for motion capture and pose estimation. This is due to their ability to detect user-defined key points without the use of labels. We present an approach for detecting robot 3D key points coordinate based on two perspectives. Two images of different perspectives are processed by the same deep neural network to detect 2D projections of key points (such as joints) associated with the robot. Triangularization is then applied to recover 3D coordinates. In addition, the existing dataset of robot critical points is obtained by simulation. It seriously impedes the advancement of this field. We establish an actual robot key points dataset and corresponding evaluation criterion. The experimental results demonstrate that our framework achieves cutting-edge detection performance. It is reusable for other robot models.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.