Potential-based shaping in model-based reinforcement learning

  • Authors:
  • John Asmuth;Michael L. Littman;Robert Zinkov

  • Affiliations:
  • RL3Laboratory, Department of Computer Science, Rutgers University, Piscataway, NJ;RL3Laboratory, Department of Computer Science, Rutgers University, Piscataway, NJ;RL3Laboratory, Department of Computer Science, Rutgers University, Piscataway, NJ

  • Venue:
  • AAAI'08 Proceedings of the 23rd national conference on Artificial intelligence - Volume 2
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

Potential-based shaping was designed as a way of introducing background knowledge into model-free reinforcement-learning algorithms. By identifying states that are likely to have high value, this approach can decrease experience complexity--the number of trials needed to find near-optimal behavior. An orthogonal way of decreasing experience complexity is to use a model-based learning approach, building and exploiting an explicit transition model. In this paper, we show how potential-based shaping can be redefined to work in the model-based setting to produce an algorithm that shares the benefits of both ideas.