Empirical evaluation methods for multiobjective reinforcement learning algorithms

  • Authors:
  • Peter Vamplew;Richard Dazeley;Adam Berry;Rustam Issabekov;Evan Dekker

  • Affiliations:
  • Graduate School of Information Technology and Mathematical Sciences, University of Ballarat, Ballarat, Australia 3353;Graduate School of Information Technology and Mathematical Sciences, University of Ballarat, Ballarat, Australia 3353;CSIRO Energy Centre, Mayfield West, Australia 2304;Graduate School of Information Technology and Mathematical Sciences, University of Ballarat, Ballarat, Australia 3353;Graduate School of Information Technology and Mathematical Sciences, University of Ballarat, Ballarat, Australia 3353

  • Venue:
  • Machine Learning
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

While a number of algorithms for multiobjective reinforcement learning have been proposed, and a small number of applications developed, there has been very little rigorous empirical evaluation of the performance and limitations of these algorithms. This paper proposes standard methods for such empirical evaluation, to act as a foundation for future comparative studies. Two classes of multiobjective reinforcement learning algorithms are identified, and appropriate evaluation metrics and methodologies are proposed for each class. A suite of benchmark problems with known Pareto fronts is described, and future extensions and implementations of this benchmark suite are discussed. The utility of the proposed evaluation methods are demonstrated via an empirical comparison of two example learning algorithms.