On the convergence of optimistic policy iteration

  • Authors:
  • John N. Tsitsiklis

  • Affiliations:
  • LIDS, Room 35-209, Massachusetts Institute of Technology, 77 Massachusetts Avenue, Cambridge, MA

  • Venue:
  • The Journal of Machine Learning Research
  • Year:
  • 2003

Quantified Score

Hi-index 0.00

Visualization

Abstract

We consider a finite-state Markov decision problem and establish the convergence of a special case of optimistic policy iteration that involves Monte Carlo estimation of Q-values, in conjunction with greedy policy selection. We provide convergence results for a number of algorithmic variations, including one that involves temporal difference learning (bootstrapping) instead of Monte Carlo estimation. We also indicate some extensions that either fail or are unlikely to go through.