Abstract

In robotics, obstacle avoidance is an essential ability for distance sensor-based robots. This type of robot has axisymmetrically distributed distance sensors to acquire obstacle distance, so the state is symmetrical. Training the control policy with a reinforcement learning method is a trend. Considering the complexity of environments, such as narrow paths and right-angle turns, robots will have a better ability if the control policy can control the steering direction and speed simultaneously. This paper proposes the multi-dimensional action control (MDAC) approach based on a reinforcement learning technique, which can be used in multiple continuous action space tasks. It adopts a hierarchical structure, which has high and low-level modules. Low-level policies output concrete actions and the high-level policy determines when to invoke low-level modules according to the environment’s features. We design robot navigation experiments with continuous action spaces to test the method’s performance. It is an end-to-end approach and can solve complex obstacle avoidance tasks in navigation.

Highlights

  • For mobile robots, obstacle avoidance is an essential and core research area for safe navigation

  • The top-down hierarchical structure is similar to the FeUdal Hierarchical reinforcement learning (HRL) used to tackle works containing subtasks

  • The high-level and low-level modules in multi-dimensional action control (MDAC) can be regarded as manager and workers

Read more

Summary

Introduction

Obstacle avoidance is an essential and core research area for safe navigation. The acquired status is distributed symmetrically, and the sensors with shorter distances are used to judge the orientation of obstacles To optimize performance, it should take lots of environmental scenarios into account to test the control policy to improve the overall behavior [1]. In the artificial intelligence domain, solving the control problems with a deep reinforcement learning (DRL) algorithm is effective since it does not need too much manual work on parameters tuning and expert knowledge. It can increase the robotics’ autonomy, and flexibility for acquiring skills [2]. Using RL algorithms to train robots has become a researching hotspot

Methods
Discussion
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.