Nonapproximability results for partially observable Markov decision processes

  • Authors:
  • Christopher Lusena;Judy Goldsmith;Martin Mundhenk

  • Affiliations:
  • Dept. of Computer Science, University of Kentucky, Lexington, KY;Dept. of Computer Science, University of Kentucky, Lexington, KY;FB IV - Informatik, Universitä Trier, Trier, Germany

  • Venue:
  • Journal of Artificial Intelligence Research
  • Year:
  • 2001

Quantified Score

Hi-index 0.00

Visualization

Abstract

We show that for several variations of partially observable Markov decision processes, polynomial-time algorithms for finding control policies are unlikely to or simply don't have guarantees of finding policies within a constant factor or a constant summand of optimal. Here "unlikely" means "unless some complexity classes collapse," where the collapses considered are P = NP, P = PSPACE, or P = EXP. Until or unless these collapses are shown to hold, any control-policy designer must choose between such performance guarantees and efficient computation.