Hierarchical solution of Markov decision processes using macro-actions

  • Authors:
  • Milos Hauskrecht;Nicolas Meuleau;Leslie Pack Kaelbling;Thomas Dean;Craig Boutilier

  • Affiliations:
  • Computer Science Department, Brown University, Providence, RI;Computer Science Department, Brown University, Providence, RI;Computer Science Department, Brown University, Providence, RI;Computer Science Department, Brown University, Providence, RI;Department of Computer Science, University of British Columbia, Vancouver, BC, Canada

  • Venue:
  • UAI'98 Proceedings of the Fourteenth conference on Uncertainty in artificial intelligence
  • Year:
  • 1998

Quantified Score

Hi-index 0.00

Visualization

Abstract

We investigate the use of temporally abstract actions, or macro-actions, in the solution of Markov decision processes. Unlike current models that combine both primitive actions and macro-actions and leave the state space unchanged, we propose a hierarchical model (using an abstract MDP) that works with macro-actions only, and that significantly reduces the size of the state space. This is achieved by treating macroactions as local policies that act in certain regions of state space, and by restricting states in the abstract MDP to those at the boundaries of regions. The abstract MDP approximates the original and can be solved more efficiently. We discuss several ways in which macro-actions can be generated to ensure good solution quality. Finally, we consider ways in which macro-actions can be reused to solve multiple, related MDPs; and we show that this can justify the computational overhead of macro-action generation.