Abstract
This paper proposes a local path planning method for the mobile robot based on the Q-Learning (QL) algorithm to improve the common problems existed in the traditional method, such as the slow convergence rate, the dilemma between exploration and exploitation, and the complex obstacles in the workplace. First, the variables of state and action are designed and discretized according to the path planning task. Then a Q-value function matrix is used to store the reinforcement value, and a reward function is constructed according to the requirements of obstacle avoidance and shortest path. To solve the balance problem of exploration and exploitation and improve the convergence speed, the e-balance strategies and the selection algorithm of actions are designed to improve the learning process of QL. After training of QL, the optimal pairs of state and action are obtained, further the optimal control rules are achieved and used to perform local path planning. In order to prevent the incomplete visiting problem for the pairs of state and action, the steering rules are appended to improve the efficiency of the path planning. Finally, the designed method has been simulated. The results show that the robot can plan an optimal or sub-optimal path while avoiding obstacles, even in a complicated environment.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.