Abstract

Planning in stochastic and partially observable environments is a central problem in artificial intelligence. To address this issue, an accurate model or a black-box simulator of the environment is usually needed in the literature. Although some recent approaches have been proposed for learning optimal behaviors under model uncertainty, prior knowledge about the environment is still required to guarantee the performance of the proposed algorithms. With the benefits of the Predictive State Representations (PSRs) approach for state representation and model prediction, in this paper, we introduce an approach for planning under partial observability with no prior domain knowledge, where an offline PSR model is firstly learned and then combined with online Monte-Carlo tree search for planning under model uncertainty. Furthermore, we also showed that with the proposed framework, the PSR models learned via other techniques, e.g., the online PSR model learning approach, can be integrated straightforwardly. By comparing with the state-of-the-art approach of planning under model uncertainty, we demonstrated the effectiveness of the proposed approaches along with the proof of their convergence.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.