Abstract

Most successes in accelerating RL incorporated internal knowledge or human intervention into the learning system such as reward shaping, transfer learning, parameter tuning, and even heuristics. These approaches could be no longer solutions to RL acceleration when internal knowledge is not available. Since the learning convergence is determined by the size of the state space where the larger the state space the slower learning might become, reducing the state space by eliminating the insignificant ones can lead to faster learning. In this paper a novel algorithm called Online State Elimination in Accelerated Reinforcement Learning (OSE-ARL) is introduced. This algorithm accelerates the RL learning performance by distinguishing insignificant states from the significant one and then eliminating them from the state space in early learning episodes. Applying OSE-ARL in grid world robot navigation shows 1.46 times faster in achieving learning convergence. This algorithm is generally applicable for other robotic task challenges or general robotics learning with large scale state space.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.