The nature of statistical learning theory
The nature of statistical learning theory
Neural Networks for Pattern Recognition
Neural Networks for Pattern Recognition
Accelerating the Convergence of Evolutionary Algorithms by Fitness Landscape Approximation
PPSN V Proceedings of the 5th International Conference on Parallel Problem Solving from Nature
Comparison Of Methods For Using Reduced Models To Speed Up Design Optimization
GECCO '02 Proceedings of the Genetic and Evolutionary Computation Conference
Using approximations to accelerate engineering design optimization
Using approximations to accelerate engineering design optimization
A comprehensive survey of fitness approximation in evolutionary computation
Soft Computing - A Fusion of Foundations, Methodologies and Applications
A framework for evolutionary optimization with approximate fitnessfunctions
IEEE Transactions on Evolutionary Computation
Multiple model regression estimation
IEEE Transactions on Neural Networks
Analytical and numerical comparisons of biogeography-based optimization and genetic algorithms
Information Sciences: an International Journal
Hi-index | 0.00 |
Real life optimization problems often require finding optimal solution to complex high dimensional, multimodal problems involving computationally very expensive fitness function evaluations.allUse of any population based iterative technique such as evolutionary algorithmall in such problem domains is thus practically prohibitive. A feasible alternative is to build surrogates or use an approximation of the actual fitness functions to be evaluated. Naturally these surrogate or meta models are order of magnitude cheaper to evaluate compared to the actual function evaluation. This paper presents two evolutionary algorithm frameworks which involve surrogate based fitness function evaluation. The first framework, namely the Dynamic Approximate Fitness based Hybrid EA (DAFHEA) model [1] reduces computation time by controlled use of meta-models (in this case approximate model generated by Support Vector Machine regression) to partially replace the actual function evaluation by approximate function evaluation. However, the underlying assumption in DAFHEA is that the training samples for the meta-model are generated from a single uniform model. This does not take into account problem domains involving uncertain environment. The second model, DAFHEA-II, an enhanced version of the original DAFHEA framework, incorporates a multiple-model based learning approach for the support vector machine approximator to handle uncertain environment [2]. Empirical evaluation results have been presented based on application of the frameworks to commonly used benchmark functions.