Qualitative analysis of partially-observable Markov decision processes

  • Authors:
  • Krishnendu Chatterjee;Laurent Doyen;Thomas A. Henzinger

  • Affiliations:
  • Institute of Science and Technology Austria;LSV, ENS Cachan & CNRS, France;Institute of Science and Technology Austria

  • Venue:
  • MFCS'10 Proceedings of the 35th international conference on Mathematical foundations of computer science
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

We study observation-based strategies for partially-observable Markov decision processes (POMDPs) with parity objectives. An observation-based strategy relies on partial information about the history of a play, namely, on the past sequence of observations. We consider qualitative analysis problems: given a POMDP with a parity objective, decide whether there exists an observation-based strategy to achieve the objective with probability 1 (almost-sure winning), or with positive probability (positive winning). Our main results are twofold. First, we present a complete picture of the computational complexity of the qualitative analysis problem for POMDPs with parity objectives and its subclasses: safety, reachability, Büchi, and coBüchi objectives. We establish several upper and lower bounds that were not known in the literature. Second, we give optimal bounds (matching upper and lower bounds) for the memory required by pure and randomized observation-based strategies for each class of objectives.