Abstract

Several approaches have been proposed to assist humans in co-manipulation and teleoperation tasks given demonstrated trajectories. However, these approaches are not applicable when the demonstrations are suboptimal or when the generalization capabilities of the learned models cannot cope with the changes in the environment. Nevertheless, in real co-manipulation and teleoperation tasks, the original demonstrations will often be suboptimal and a learning system must be able to cope with new situations. This paper presents a reinforcement learning algorithm that can be applied to such problems. The proposed algorithm is initialized with a probability distribution of demonstrated trajectories and is based on the concept of relevance functions. We show in this paper how the relevance of trajectory parameters to optimization objectives is connected with the concept of Pearson correlation. First, we demonstrate the efficacy of our algorithm by addressing the assisted teleoperation of an object in a static virtual environment. Afterward, we extend this algorithm to deal with dynamic environments by utilizing Gaussian Process regression. The full framework is applied to make a point particle and a 7-DoF robot arm autonomously adapt their movements to changes in the environment as well as to assist the teleoperation of a 7-DoF robot arm in a dynamic environment.

Highlights

  • Learning from demonstrations is a promising approach toward human-robot co-manipulation and teleoperation

  • We extend PRO with Gaussian Processes (GP) regression to cope with dynamic environments

  • To adapt Probabilistic Movement Primitives (ProMPs) on the fly to changes in the environment, our learning system must be able to compute these ProMPs quickly. To deal with this challenge, we propose using Gaussian Process (GP) regression to map variables describing the environment to mean vector μw and covariance matrix w of a ProMP

Read more

Summary

INTRODUCTION

Learning from demonstrations is a promising approach toward human-robot co-manipulation and teleoperation. Our work contributes to this field by providing a new reinforcement learning algorithm, Pearson-Correlation-Based Relevance Weighted Policy Optimization (PRO), to improve upon demonstrated trajectories when these are suboptimal or when solutions to new situations must be found. These trajectories need to be optimized with respect to objectives, such as minimizing distances to via points, keeping a certain minimum distance from obstacles, achieving minimal length, minimal jerk, etc. The new algorithm presented in this paper, PRO, is based on the insight that the Pearson correlation coefficient (Benesty et al, 2009) can be used to determine how each trajectory parameter influences each objective It does not require designing basis functions for the relevance. Excerpts of this work have been accepted for presentation at Ewerton et al (2019)

RELATED WORK
Relevance Functions
Optimization of Trajectory
ONLINE ADAPTATION OF TRAJECTORY DISTRIBUTIONS
6: Sample trajectory parameters w from N
EXPERIMENTS
Assisted Teleoperation of a Virtual Object
Adaptation in Dynamic
Adaptation in Dynamic Environments—Autonomous Robot Arm
Teleoperation of a Robot Arm in a Dynamic Environment
CONCLUSION AND FUTURE WORK
DATA AVAILABILITY STATEMENT
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call