Extreme compass and dynamic multi-armed bandits for adaptive operator selection

  • Authors:
  • Jorge Maturana;Álvaro Fialho;Frédéric Saubion;Marc Schoenauer;Michèle Sebag

  • Affiliations:
  • LERIA laboratory, Université d'Angers, Angers, France;Microsoft Research, INRIA Joint Centre, Orsay, France;LERIA laboratory, Université d'Angers, Angers, France;Microsoft Research, INRIA Joint Centre, Orsay, France and INRIA Saclay, Île-de-France and LRI, UMR, CNRS, Orsay, France;Microsoft Research, INRIA Joint Centre, Orsay, France and INRIA Saclay, Île-de-France and LRI, UMR, CNRS, Orsay, France

  • Venue:
  • CEC'09 Proceedings of the Eleventh conference on Congress on Evolutionary Computation
  • Year:
  • 2009

Quantified Score

Hi-index 0.01

Visualization

Abstract

The goal of Adaptive Operator Selection is the on-line control of the choice of variation operators within Evolutionary Algorithms. The control process is based on two main components, the credit assignment, that defines the reward that will be used to evaluate the quality of an operator after it has been applied, and the operator selection mechanism, that selects one operator based on some operators qualities. Two previously developed Adaptive Operator Selection methods are combined here: Compass evaluates the performance of operators by considering not only the fitness improvements from parent to offspring, but also the way they modify the diversity of the population, and their execution time; Dynamic Multi-Armed Bandit proposes a selection strategy based on the well-known UCB algorithm, achieving a compromise between exploitation and exploration, while nevertheless quickly adapting to changes. Tests with the proposed method, called ExCoDyMAB, are carried out using several hard instances of the Satisfiability problem (SAT). Results show the good synergetic effect of combining both approaches.