Dynamic Multi-Armed Bandits and Extreme Value-Based Rewards for Adaptive Operator Selection in Evolutionary Algorithms

  • Authors:
  • Álvaro Fialho;Luis Costa;Marc Schoenauer;Michèle Sebag

  • Affiliations:
  • Microsoft Research --- INRIA Joint Centre, Orsay, France;TAO team, INRIA Saclay --- Île-de-France & LRI (UMR CNRS 8623), Orsay, France;Microsoft Research --- INRIA Joint Centre, Orsay, France and TAO team, INRIA Saclay --- Île-de-France & LRI (UMR CNRS 8623), Orsay, France;Microsoft Research --- INRIA Joint Centre, Orsay, France and TAO team, INRIA Saclay --- Île-de-France & LRI (UMR CNRS 8623), Orsay, France

  • Venue:
  • Learning and Intelligent Optimization
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

The performance of many efficient algorithms critically depends on the tuning of their parameters, which on turn depends on the problem at hand. For example, the performance of Evolutionary Algorithms critically depends on the judicious setting of the operator rates. The Adaptive Operator Selection (AOS) heuristic that is proposed here rewards each operator based on the extreme value of the fitness improvement lately incurred by this operator, and uses a Multi-Armed Bandit (MAB) selection process based on those rewards to choose which operator to apply next. This Extreme-based Multi-Armed Bandit approach is experimentally validated against the Average-based MAB method, and is shown to outperform previously published methods, whether using a classical Average-based rewarding technique or the same Extreme-based mechanism. The validation test suite includes the easy One-Max problem and a family of hard problems known as "Long k-paths".