Structured reachability analysis for Markov decision processes

  • Authors:
  • Craig Boutilier;Ronen I. Brafman;Christopher Geib

  • Affiliations:
  • Department of Computer Science, University of British Columbia, Vancouver, BC, Canada;Department of Math and CS, Ben-Gurion University, Beer Sheva, Israel;Honeywell Technology Center, Minneapolis, MN

  • Venue:
  • UAI'98 Proceedings of the Fourteenth conference on Uncertainty in artificial intelligence
  • Year:
  • 1998

Quantified Score

Hi-index 0.00

Visualization

Abstract

Recent research in decision theoretic planning has focussed on making the solution of Markov decision processes (MDPs) more feasible. We develop a family of algorithms for structured reachability analysis of MDPs that are suitable when an initial state (or set of states) is known. Usin compact, structured representations of MDPs (e.g., Bayesian networks), our methods, which vary in the tradeoff between complexity and accurac roduce structured descriptions of (estimated) reacpagle states that can be used to eliminate variables oy variable values from the problem description, reducing the size of the MDP and making it easier to solve. One contribution of our work is the extension of ideas from GRAPHPLAN to deal with the distributed nature of action reoresentations typically embodied within Bayes nets and the problem of correlated action effects. We also demonstrate that our algorithm can be made more complete by using k-ary constraints instead of binary constraints. Another contribution is the illustration of how the compact representation of reachability constraints can be exploited by several existing (exact and approximate) abstraction algorithms for MDPs.