Piecewise linear value function approximation for factored MDPs

  • Authors:
  • Pascal Poupart;Craig Boutilier;Relu Patrascu;Dale Schuurmans

  • Affiliations:
  • Dept. of Computer Science, University of Toronto, Toronto, ON, M5S 3H5;Dept. of Computer Science, University of Toronto, Toronto, ON, M5S 3H5;Department of Computer Science, University of Waterloo, Waterloo, ON, N2L 3G1;Department of Computer Science, University of Waterloo, Waterloo, ON, N2L 3G1

  • Venue:
  • Eighteenth national conference on Artificial intelligence
  • Year:
  • 2002

Quantified Score

Hi-index 0.00

Visualization

Abstract

A number of proposals have been put forth in recent years for the solution of Markov decision processes (MDPs) whose state (and sometimes action) spaces are factored. One recent class of methods involves linear value function approximation, where the optimal value function is assumed to be a linear combination of some set of basis functions, with the aim of finding suitable weights. While sophisticated techniques have been developed for finding the best approximation within this constrained space, few methods have been proposed for choosing a suitable basis set, or modifying it if solution quality is found wanting. We propose a general framework, and specific proposals, that address both of these questions. In particular, we examine weakly coupled MDPs where a number of subtasks can be viewed independently modulo resource constraints. We then describe methods for constructing a piecewise linear combination of the subtask value functions, using greedy decision tree techniques. We argue that this architecture is suitable for many types of MDPs whose combinatorics are determined largely by the existence multiple conflicting objectives.