Efficient Global Optimization of Expensive Black-Box Functions
Journal of Global Optimization
Genetic Search with Approximate Function Evaluation
Proceedings of the 1st International Conference on Genetic Algorithms
Metamodel-Assisted Evolution Strategies
PPSN VII Proceedings of the 7th International Conference on Parallel Problem Solving from Nature
A Comparison of Approximation Modeling Techniques: Polynomial Versus Interpolating Models
A Comparison of Approximation Modeling Techniques: Polynomial Versus Interpolating Models
A comprehensive survey of fitness approximation in evolutionary computation
Soft Computing - A Fusion of Foundations, Methodologies and Applications
Proceedings of the 10th international conference on Parallel Problem Solving from Nature: PPSN X
A systems approach to evolutionary multiobjective structural optimization and beyond
IEEE Computational Intelligence Magazine
Generalizing surrogate-assisted evolutionary computation
IEEE Transactions on Evolutionary Computation
Accelerating evolutionary algorithms with Gaussian process fitness function models
IEEE Transactions on Systems, Man, and Cybernetics, Part C: Applications and Reviews
A framework for evolutionary optimization with approximate fitnessfunctions
IEEE Transactions on Evolutionary Computation
Single- and multiobjective evolutionary optimization assisted by Gaussian random field metamodels
IEEE Transactions on Evolutionary Computation
Hi-index | 0.00 |
Surrogate models used in evolutionary algorithms (EAs) aim to reduce computationally expensive objective function evaluations. However, low-quality surrogates may mislead EAs and as a result, surrogate-assisted EAs may fail to locate the global optimum. Among various machine learning models for surrogates, Gaussian Process (GP) models have shown to be effective as GP models are able to provide fitness estimation as well as a confidence level. One weakness of GP models is that the computational cost for training increases rapidly as the number of training samples increases. To reduce the computational cost for training, here we propose to adopt an ensemble of local Gaussian Process models. Different from independent local Gaussian Process models, local Gaussian Process models share the same model parameters. Then the performance of the covariance matrix adaptation evolution strategy (CMA-ES) assisted by an ensemble of local Gaussian Process models with five different sampling strategies is compared. Experiments on eight benchmark functions demonstrate that ensembles of local Gaussian Process models can provide reliable fitness prediction and uncertainty estimation. Among the compared strategies, the clustering technique using the lower confidence bound sampling strategy exhibits the best global search performance.