A Generalized Kalman Filter for Fixed Point Approximation and Efficient Temporal-Difference Learning

  • Authors:
  • David Choi;Benjamin Roy

  • Affiliations:
  • Lincoln Laboratory, Massachusetts Institue of Technology, Lexington, USA 02420-9108;Departments of Management Science and Engineering and Electrical Engineering, Stanford University, Stanford, USA 94305

  • Venue:
  • Discrete Event Dynamic Systems
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

The traditional Kalman filter can be viewed as a recursive stochastic algorithm that approximates an unknown function via a linear combination of prespecified basis functions given a sequence of noisy samples. In this paper, we generalize the algorithm to one that approximates the fixed point of an operator that is known to be a Euclidean norm contraction. Instead of noisy samples of the desired fixed point, the algorithm updates parameters based on noisy samples of functions generated by application of the operator, in the spirit of Robbins---Monro stochastic approximation. The algorithm is motivated by temporal-difference learning, and our developments lead to a possibly more efficient variant of temporal-difference learning. We establish convergence of the algorithm and explore efficiency gains through computational experiments involving optimal stopping and queueing problems.