Point-Based Methods for Model Checking in Partially Observable Markov Decision Processes

Authors

  • Maxime Bouton Stanford University
  • Jana Tumova KTH Royal Institute of Technology
  • Mykel J. Kochenderfer Stanford University

DOI:

https://doi.org/10.1609/aaai.v34i06.6563

Abstract

Autonomous systems are often required to operate in partially observable environments. They must reliably execute a specified objective even with incomplete information about the state of the environment. We propose a methodology to synthesize policies that satisfy a linear temporal logic formula in a partially observable Markov decision process (POMDP). By formulating a planning problem, we show how to use point-based value iteration methods to efficiently approximate the maximum probability of satisfying a desired logical formula and compute the associated belief state policy. We demonstrate that our method scales to large POMDP domains and provides strong bounds on the performance of the resulting policy.

Downloads

Published

2020-04-03

How to Cite

Bouton, M., Tumova, J., & Kochenderfer, M. J. (2020). Point-Based Methods for Model Checking in Partially Observable Markov Decision Processes. Proceedings of the AAAI Conference on Artificial Intelligence, 34(06), 10061-10068. https://doi.org/10.1609/aaai.v34i06.6563

Issue

Section

AAAI Technical Track: Reasoning under Uncertainty