An Overview of MAXQ Hierarchical Reinforcement Learning

  • Authors:
  • Thomas G. Dietterich

  • Affiliations:
  • -

  • Venue:
  • SARA '02 Proceedings of the 4th International Symposium on Abstraction, Reformulation, and Approximation
  • Year:
  • 2000

Quantified Score

Hi-index 0.00

Visualization

Abstract

Reinforcement learning addresses the problem of learning optimal policies for sequential decision-making problems involving stochastic operators and numerical reward functions rather than the more traditional deterministic operators and logical goal predicates. In many ways, reinforcement learning research is recapitulating the development of classical research in planning and problem solving. After studying the problem of solving "flat" problem spaces, researchers have recently turned their attention to hierarchical methods that incorporate subroutines and state abstractions. This paper gives an overview of the MAXQ value function decomposition and its support for state abstraction and action abstraction.