Abstract
The morphology of a robot is typically assumed to be known, and data from external measuring devices are used mainly for its kinematic calibration. In contrast, we take an agent-centric perspective and ponder the vaguely explored question of whether a robot could learn elements of its morphology by itself, relying on minimal prior knowledge and depending only on unorganized proprioceptive signals. To answer this question, we propose a mutual information–based representation of the relationships between the proprioceptive signals of a robot, which we call proprioceptive information graphs (π-graphs). Leveraging the fact that the information structure of the sensorimotor apparatus is dependent on the embodiment of the robot, we use the π-graph to look for pairwise signal relationships that reflect the underlying kinematic first-order principles applicable to the robot’s structure. In our discussion, we show that analysis of the π-graph leads to the inference of two fundamental elements of the robot morphology: its mechanical topology and corresponding kinematic description, that is, the location and orientation of the robot’s joints. Results from a robot manipulator, a hexapod, and a humanoid robot show that the correct topology and kinematic description can be effectively inferred from their π-graph either offline or online, regardless of the number of links and body configuration.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.