Random Sampling of States in Dynamic Programming

  • Authors:
  • C. G. Atkeson;B. J. Stephens

  • Affiliations:
  • Robot. Inst., Carnegie Mellon Univ., Pittsburgh, PA;-

  • Venue:
  • IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

We combine three threads of research on approximate dynamic programming: sparse random sampling of states, value function and policy approximation using local models, and using local trajectory optimizers to globally optimize a policy and associated value function. Our focus is on finding steady-state policies for deterministic time-invariant discrete time control problems with continuous states and actions often found in robotics. In this paper, we describe our approach and provide initial results on several simulated robotics problems.