Abstract
Continuously learning new tasks using high-level ideas or knowledge is a key capability of humans. In this paper, we propose lifelong reinforcement learning with sequential linear temporal logic formulas and reward machines (LSRM), which enables an agent to leverage previously learned knowledge to accelerate the learning of logically specified tasks. For a more flexible specification of tasks, we first introduce sequential linear temporal logic (SLTL), which is a supplement to the existing linear temporal logic (LTL) formal language. We then utilize reward machines (RMs) to exploit structural reward functions for tasks encoded with high-level events, and propose an automatic extension of RMs and efficient knowledge transfer over tasks for continuous lifelong learning. Experimental results show that LSRM outperforms methods that learn the target tasks from scratch by taking advantage of the task decomposition using SLTL and the knowledge transfer over RMs during the lifelong learning process.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.