Stochastic iterative dynamic programming: a Monte Carlo approach to dual control

  • Authors:
  • Adrian M. Thompson;William R. Cluett

  • Affiliations:
  • Department of Chemical Engineering and Applied Chemistry, University of Toronto, Toronto, Ont., Canada M5S 3E5;Department of Chemical Engineering and Applied Chemistry, University of Toronto, Toronto, Ont., Canada M5S 3E5

  • Venue:
  • Automatica (Journal of IFAC)
  • Year:
  • 2005

Quantified Score

Hi-index 22.14

Visualization

Abstract

Practical exploitation of optimal dual control (ODC) theory continues to be hindered by the difficulties involved in numerically solving the associated stochastic dynamic programming (SDPs) problems. In particular, high-dimensional hyper-states coupled with the nesting of optimizations and integrations within these SDP problems render their exact numerical solution computationally prohibitive. This paper presents a new stochastic dynamic programming algorithm that uses a Monte Carlo approach to circumvent the need for numerical integration, thereby dramatically reducing computational requirements. Also, being a generalization of iterative dynamic programming (IDP) to the stochastic domain, the new algorithm exhibits reduced sensitivity to the hyper-state dimension and, consequently, is particularly well suited to solution of ODC problems. A convergence analysis of the new algorithm is provided, and its benefits are illustrated on the problem of ODC of an integrator with unknown gain, originally presented by Astrom and Helmersson (Computers and Mathematics with Applications 12A (1986) 653-662).