Markov decision processes with multiple long-run average objectives

  • Authors:
  • Krishnendu Chatterjee

  • Affiliations:
  • UC Berkeley

  • Venue:
  • FSTTCS'07 Proceedings of the 27th international conference on Foundations of software technology and theoretical computer science
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

We consider Markov decision processes (MDPs) with multiple long-run average objectives. Such MDPs occur in design problems where one wishes to simultaneously optimize several criteria, for example, latency and power. The possible trade-offs between the different objectives are characterized by the Pareto curve. We show that every Pareto optimal point can be Ɛ-approximated by a memoryless strategy, for all Ɛ 0. In contrast to the single-objective case, the memoryless strategy may require randomization. We show that the Pareto curve can be approximated (a) in polynomial time in the size of the MDP for irreducible MDPs; and (b) in polynomial space in the size of the MDP for all MDPs. Additionally, we study the problem if a given value vector is realizable by any strategy, and show that it can be decided in polynomial time for irreducible MDPs and in NP for all MDPs. These results provide algorithms for design exploration in MDP models with multiple long-run average objectives.