Abstract

Markov decision processes are of major interest in the planning community as well as in the model checking community. But in spite of the similarity in the considered formal models, the development of new techniques and methods happened largely independently in both communities. This work is intended as a beginning to unite the two research branches. We consider goal-reachability analysis as a common basis between both communities. The core of this paper is the translation from Jani, an overarching input language for quantitative model checkers, into the probabilistic planning domain definition language (PPDDL), and vice versa from PPDDL into Jani. These translations allow the creation of an overarching benchmark collection, including existing case studies from the model checking community, as well as benchmarks from the international probabilistic planning competitions (IPPC). We use this benchmark set as a basis for an extensive empirical comparison of various approaches from the model checking community, variants of value iteration, and MDP heuristic search algorithms developed by the AI planning community. On a per benchmark domain basis, techniques from one community can achieve state-ofthe-art performance in benchmarks of the other community. Across all benchmark domains of one community, the performance comparison is however in favor of the solvers and algorithms of that particular community. Reasons are the design of the benchmarks, as well as tool-related limitations. Our translation methods and benchmark collection foster crossfertilization between both communities, pointing out specific opportunities for widening the scope of solvers to different kinds of models, as well as for exchanging and adopting algorithms across communities.

Highlights

  • Running systems in the realworld adds uncertainty to the system’s execution from all kinds of sources

  • In own work motivated by our observations here, we developed a preliminary adaptation of Markov decision processes (MDPs) heuristic search with FRET in the Modest Toolset

  • As has been shown in previous works (Steinmetz et al, 2016), and as we will confirm in Section 7.2 and 8, heuristic search can be useful even when ran without an informative heuristic

Read more

Summary

Introduction

Running systems (be it purely software based, or with a physical component) in the realworld adds uncertainty to the system’s execution from all kinds of sources. This uncertainty can for instance come from the environment itself (e.g., external events), the possibility of failures (e.g., a robot hand might fail to grasp), or even the intrinsic probabilistic nature of actions (e.g., tossing a coin). Given an MDP, in planning one is usually interested in finding a way to reach some predefined goal starting from an initial state. We are interested in MaxProb (Kolobov, Mausam, Weld, & Geffner, 2011; Teichteil-Konigsbuch, 2012; Trevizan, Teichteil-Knigsbuch, & Thiebaux, 2017), which is finding the policy whose execution leads to the satisfaction of the goal with maximal possible probability

Results
Discussion
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.