Using background knowledge to speed reinforcement learning in physical agents

  • Authors:
  • Daniel Shapiro;Pat Langley;Ross Shachter

  • Affiliations:
  • Department of Management Science and Engineering, Stanford University, Stanford, CA;Institute for the Study of Learning and Expertise, 2164 Staunton Court, Palo Alto, CA;Department of Management Science and Engineering, Stanford University, Stanford, CA

  • Venue:
  • Proceedings of the fifth international conference on Autonomous agents
  • Year:
  • 2001

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper describes Icarus, an agent architecture that embeds a hierarchical reinforcement learning algorithm within a language for specifying agent behavior. An Icarus program expresses an approximately correct theory about how to behave with options at varying levels of detail, while the Icarus agent determines the best options by learning from experience. We describe Icarus and its learning algorithm, then report on two experiments in a vehicle control domain. The first examines the benefit of new distinctions about state, whereas the second explores the impact of added plan structure. We show that background knowledge increases learning rate and asymptotic performance, and decreases plan size by three orders of magnitude, relative to the typical formulation of the learning problem in our test domain.