Bayesian real-time dynamic programming

  • Authors:
  • Scott Sanner;Robby Goetschalckx;Kurt Driessens;Guy Shani

  • Affiliations:
  • SML Group, National ICT Australia, Canberra, Australia;Department of Computer Science, Catholic University of Leuven, Heverlee, Belgium;Department of Computer Science, Catholic University of Leuven, Heverlee, Belgium;MLAS Group, Microsoft Research, Redmond, WA

  • Venue:
  • IJCAI'09 Proceedings of the 21st international jont conference on Artifical intelligence
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Real-time dynamic programming (RTDP) solves Markov decision processes (MDPs) when the initial state is restricted, by focusing dynamic programming on the envelope of states reachable from an initial state set. RTDP often provides performance guarantees without visiting the entire state space. Building on RTDP, recent work has sought to improve its efficiency through various optimizations, including maintaining upper and lower bounds to both govern trial termination and prioritize state exploration. In this work, we take a Bayesian perspective on these upper and lower bounds and use a value of perfect information (VPI) analysis to govern trial termination and exploration in a novel algorithm we call VPI-RTDP. VPI-RTDP leads to an improvement over state-of-the-art RTDP methods, empirically yielding up to a three-fold reduction in the amount of time and number of visited states required to achieve comparable policy performance.