Abstract

Current robots can perform repetitive tasks well, but are constrained to environment and task variations. Teaching a robot by demonstration is a powerful approach to solve the problem. The learning methods using large sensory and joint state information are extremely difficult to efficiently learn the demonstration policy. This article proposes a learning-by-imitation approach that learns demonstration policy for robotic manipulation skill acquisition from <italic xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">what-where-how</i> interaction data. The method can improve the robotic adaptability to the environment and tasks with fewer training inputs. RGB-D image interaction demonstration is used. At each time step, we interact with an object and select a high-level action. The demonstration is formed through multistep interactions. An imitation learning architecture (OPLN) consisting of the objects list network (OLN) and policy learning network (PLN) is proposed. OLN and PLN are constructed, respectively, with long short-term memory (LSTM) neural networks. OLN learns objects sequence feature extracted from demonstration data while PLN learns policy. An action and a target object are obtained as outputs to control robot’s manipulation. The experiments show that the Block Stacking skill and Pick and Place skill can be successfully acquired, and the method can adapt to environment variations and generalize to similar tasks.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call