Abstract
Deep reinforcement learning research in a single-task environment has made remarkable achievements. However, it is often plagued by catastrophic forgetting, prohibitively low sample efficiency and lack of scalability problems when facing multi-task environment. To solve these issues, a Hierarchical Lifelong Reinforcement Learning framework (HLifeRL) is proposed to enhance the ability of agents to deal with a sequence of tasks in the way of skill discovery (we treat option as low-level skill in this paper) and hierarchical policy. HLifeRL can automatically extract task-related knowledge without any human intervention or priori knowledge. Moreover, with the help of a scalable library and the master policy, we can flexibly combine various skills to complete multiple tasks in the form of call-and-return. The experimental results show that HLifeRL can accelerate the speed of single-task training and deliver remarkable stability along with scalability in a lifelong setting environment.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: Journal of King Saud University - Computer and Information Sciences
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.