An evolution strategy assisted by an ensemble of local Gaussian process models

  • Authors:
  • Jianfeng Lu;Bin Li;Yaochu Jin

  • Affiliations:
  • University of Science and Technology of China, Hefei, China;University of Science and Technology of China, Hefei, China;University of Surrey, Guildford, United Kingdom

  • Venue:
  • Proceedings of the 15th annual conference on Genetic and evolutionary computation
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Surrogate models used in evolutionary algorithms (EAs) aim to reduce computationally expensive objective function evaluations. However, low-quality surrogates may mislead EAs and as a result, surrogate-assisted EAs may fail to locate the global optimum. Among various machine learning models for surrogates, Gaussian Process (GP) models have shown to be effective as GP models are able to provide fitness estimation as well as a confidence level. One weakness of GP models is that the computational cost for training increases rapidly as the number of training samples increases. To reduce the computational cost for training, here we propose to adopt an ensemble of local Gaussian Process models. Different from independent local Gaussian Process models, local Gaussian Process models share the same model parameters. Then the performance of the covariance matrix adaptation evolution strategy (CMA-ES) assisted by an ensemble of local Gaussian Process models with five different sampling strategies is compared. Experiments on eight benchmark functions demonstrate that ensembles of local Gaussian Process models can provide reliable fitness prediction and uncertainty estimation. Among the compared strategies, the clustering technique using the lower confidence bound sampling strategy exhibits the best global search performance.