Abstract
In this letter, we proposea novel vision-based grasp system for closed-loop 6-degrees of freedom grasping of unknown objects in cluttered environments. The key factor in our system is that we make the most of a geometry-aware scene representation based on a truncated signed distance function (TSDF) volume, which can handle multi-view observations from the vision sensor, provide comprehensive spatial information for the grasp pose detector, and allow collision checking to achieve a collision-free grasp pose. To eliminate the large computational burden caused by the volume-based data, a lightweight volume-point network (VPN) equipped with the Marching Cubes algorithm is proposed. This network predicts the point-wise grasp qualities for all the candidates in a single feed-forward operation with real-time performance, enabling the system to perform closed-loop grasping. Furthermore, a grasp pose refinement module is integrated to predict the pose residual based on the SDF observation of the gripper state in the TSDF volume. Extensive experiments show that the proposed method can achieve collision-free grasp detection with an efficiency of more than 30 Hz on a high-resolution volume. Furthermore, the model trained on only synthetic data achieves a 90.9% grasp success rate on cluttered real-world scenes, significantly outperforming other baseline methods. The supplementary materials are available at <uri>https://sites.google.com/view/vpn-icra2022</uri>
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.