Stochastic search using the natural gradient

  • Authors:
  • Sun Yi;Daan Wierstra;Tom Schaul;Jürgen Schmidhuber

  • Affiliations:
  • IDSIA, Manno, Switzerland;IDSIA, Manno, Switzerland;IDSIA, Manno, Switzerland;IDSIA, Manno, Switzerland

  • Venue:
  • ICML '09 Proceedings of the 26th Annual International Conference on Machine Learning
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

To optimize unknown 'fitness' functions, we present Natural Evolution Strategies, a novel algorithm that constitutes a principled alternative to standard stochastic search methods. It maintains a multinormal distribution on the set of solution candidates. The Natural Gradient is used to update the distribution's parameters in the direction of higher expected fitness, by efficiently calculating the inverse of the exact Fisher information matrix whereas previous methods had to use approximations. Other novel aspects of our method include optimal fitness baselines and importance mixing, a procedure adjusting batches with minimal numbers of fitness evaluations. The algorithm yields competitive results on a number of benchmarks.