Abstract
We present a novel hierarchical control framework that unifies our previous work on tactile-servoing with visual-servoing approaches to allow for robust manipulation and exploration of unknown objects, including — but not limited to — robust grasping, online grasp optimization, in-hand manipulation, and exploration of object surfaces. The framework is divided into three layers: a joint-level layer, a tactile servoing layer, and a visual servoing layer. While the middle layer provides blind surface exploration skills, maintaining desired contact patterns, the visual layer monitors and controls the actual object pose providing high-level fingertip motion commands that are merged with the tactile-servoing control commands. We illustrate the versatility of the proposed framework using a series of manipulation actions performed with two KUKA LWR arms equipped with a tactile sensor array as a sensitive fingertip. The two considered objects are unknown to the robot, i.e. neither shape nor friction properties are available.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.