Abstract

Compared to traditional data-driven learning methods, recently developed deep reinforcement learning (DRL) approaches can be employed to train robot agents to obtain control policies with appealing performance. However, learning control policies for real-world robots through DRL is costly and cumbersome. A promising alternative is to train policies in simulated environments and transfer the learned policies to real-world scenarios. Unfortunately, due to the reality gap between simulated and real-world environments, the policies learned in simulated environments often cannot be generalized well to the real world. Bridging the reality gap is still a challenging problem. In this paper, we propose a novel real–sim–real (RSR) transfer method that includes a real-to-sim training phase and a sim-to-real inference phase. In the real-to-sim training phase, a task-relevant simulated environment is constructed based on semantic information of the real-world scenario and coordinate transformation, and then a policy is trained with the DRL method in the built simulated environment. In the sim-to-real inference phase, the learned policy is directly applied to control the robot in real-world scenarios without any real-world data. Experimental results in two different robot control tasks show that the proposed RSR method can train skill policies with high generalization performance and significantly low training costs.

Highlights

  • Over the past decades, robots have been gradually applied in various fields, with the expectation of completing more control tasks for human beings

  • We study the possibility of directly transferring the policies trained in simulated environments to the real world with high generation capability and low costs

  • We proposed a novel real–sim–real (RSR) transfer method for control policy learning in real-world robots

Read more

Summary

Introduction

Robots have been gradually applied in various fields, with the expectation of completing more control tasks for human beings. One promising method is to train control policies in simulated environments where data generation is safe, convenient, and linvolves a ower cost, and to transfer the learned policies to the real world. To train control policies for real-world robots with high generalization capability, and to greatly reduce the training cost, in this work we propose a real–sim–real (RSR) transfer method that includes a real-to-sim training phase and a sim-to-real inference phase. Experimental results show that the proposed RSR method can train control policies for real-world robots with promising generalization performance and significantly low training costs. The proposed method automatically constructs a task-relevant simulated environment for policy learning based on semantic information of real-world working scenarios and coordinate transformation, which avoids the challenging problem of manually creating the simulated environments with high fidelity, endowing the policy learning process with high efficiency.

Robot Control Policy Learning
Sim-to-Real Transfer
Method
Generating a Simulated Environment
Policy Network
Policy Training
Deploying the Trained Policy
Performance Evaluation
Experiments and Results
Semantic Segmentation of Robot Working Scenarios
Policy Learning
Methods
Conclusions
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call