Improving optimistic exploration in model-free reinforcement learning

  • Authors:
  • Marek Grześ;Daniel Kudenko

  • Affiliations:
  • Department of Computer Science, University of York, Heslington, York, United Kingdom;Department of Computer Science, University of York, Heslington, York, United Kingdom

  • Venue:
  • ICANNGA'09 Proceedings of the 9th international conference on Adaptive and natural computing algorithms
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

The key problem in reinforcement learning is the explorationexploitation tradeoff. An optimistic initialisation of the value function is a popular RL strategy. The problem of this approach is that the algorithm may have relatively low performance after many episodes of learning. In this paper, two extensions to standard optimistic exploration are proposed. The first one is based on different initialisation of the value function of goal states. The second one which builds on the previous idea explicitly separates propagation of low and high values in the state space. Proposed extensions show improvement in empirical comparisons with basic optimistic initialisation. Additionally, they improve anytime performance and help on domains where learning takes place on the subspace of the large state space, that is, where the standard optimistic approach faces more difficulties.