Abstract

In this paper, we proposed a novel Hybrid Reinforcement Learning framework to maintain the stability of a biped robot (NAO) while it is walking on static and dynamic platforms. The reinforcement learning framework consists of the Model-based off-line Estimator, the Actor Network Pre-training scheme, and the Mode-free on-line optimizer. We proposed the Hierarchical Gaussian Processes as the Mode-based Estimator to predict a rough model of the system and to obtain the initial control input. Then, the initial control input is employed to pre-train the Actor Network by using the initial control input. Finally, a model-free optimizer based on Deep Deterministic Policy Gradient framework is introduced to fine tune the Actor Network and to generate the best actions. The proposed reinforcement learning framework not only successfully avoids the distribution mismatch problem while combining model-based scheme with model-free structure, but also improves the sample efficiency for the on-line learning procedure. Simulation results show that the proposed Hybrid Reinforcement Learning mechanism enables the NAO robot to maintain balance while walking on static and dynamic platforms. The robustness of the learned controllers in adapting to platforms with different angles, different magnitudes, and different frequencies is tested.

Highlights

  • The biped robot is a unique robotic system that has two legs, and it is able to perform static and dynamic postures or walking gait as humans

  • The result of Hierarchical Gaussian Processes (HGP) provided a rough model of the system and generated an initial control input, after which the Actor Critic Pre-training scheme was proposed to pre-train the Actor Network by considering the obtained initial control input as the ground-truth

  • The Pretraining procedure bridged the gap between the HGP based mode-based estimator and the Deep Deterministic Policy Gradient (DDPG) based mode-free optimizer

Read more

Summary

INTRODUCTION

The biped robot is a unique robotic system that has two legs, and it is able to perform static and dynamic postures or walking gait as humans. Lin et al [16] successfully applied Q-Learning, a model-free RL framework, to a NAO robot as the joint controller to maintain the stability of walking on flat surface, inclined and declined surface, and a seesaw They proved that the proposed Q-learning scheme was able to imitate human motions captured from cameras while still maintaining the balance without falling down [17]. By introducing the pre-training procedure, the proposed control method successfully bridges the gap between model-based RL and modelfree RL, where the transition from learning the model to obtaining the best action has been smoothly completed It avoids the distribution mismatch problem while integrating the model-based scheme into the model-free framework.

FORMULATION OF THE PROBLEM
HIERARCHICAL GAUSSIAN PROCESSES
MODEL FREE FINE TUNING
19: End For
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call