Reinforcement Learning with Reward Shaping and Mixed Resolution Function Approximation

  • Authors:
  • Marek Grzes;Daniel Kudenko

  • Affiliations:
  • University of York, UK;University of York, UK

  • Venue:
  • International Journal of Agent Technologies and Systems
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

A crucial trade-off is involved in the design process when function approximation is used in reinforcement learning. Ideally the chosen representation should allow representing as close as possible an approximation of the value function. However, the more expressive the representation the more training data is needed because the space of candidate hypotheses is bigger. A less expressive representation has a smaller hypotheses space and a good candidate can be found faster. The core idea of this paper is the use of a mixed resolution function approximation, that is, the use of a less expressive function approximation to provide useful guidance during learning, and the use of a more expressive function approximation to obtain a final result of high quality. A major question is how to combine the two representations. Two approaches are proposed and evaluated empirically.