Evolutionary algorithms in theory and practice: evolution strategies, evolutionary programming, genetic algorithms
Swarm intelligence
Genetic Algorithms in Search, Optimization and Machine Learning
Genetic Algorithms in Search, Optimization and Machine Learning
Parameter Selection in Particle Swarm Optimization
EP '98 Proceedings of the 7th International Conference on Evolutionary Programming VII
Comparison between Genetic Algorithms and Particle Swarm Optimization
EP '98 Proceedings of the 7th International Conference on Evolutionary Programming VII
Evolutionary Optimization Versus Particle Swarm Optimization: Philosophy and Performance Differences
EP '98 Proceedings of the 7th International Conference on Evolutionary Programming VII
Population structure and particle swarm performance
CEC '02 Proceedings of the Evolutionary Computation on 2002. CEC '02. Proceedings of the 2002 Congress - Volume 02
Population structure and particle swarm performance
CEC '02 Proceedings of the Evolutionary Computation on 2002. CEC '02. Proceedings of the 2002 Congress - Volume 02
Self-organizing hierarchical particle swarm optimizer with time-varying acceleration coefficients
IEEE Transactions on Evolutionary Computation
A hybrid attractive and repulsive particle swarm optimization based on gradient search
ICIC'13 Proceedings of the 9th international conference on Intelligent Computing Theories and Technology
A parallel hybrid optimization algorithm for fitting interatomic potentials
Applied Soft Computing
Hi-index | 0.00 |
Stochastic optimization algorithms like genetic algorithms (GAs) and particle swarm optimization (PSO) algorithms perform global optimization but waste computational effort by doing a random search. On the other hand deterministic algorithms like gradient descent converge rapidly but may get stuck in local minima of multimodal functions. Thus, an approach that combines the strengths of stochastic and deterministic optimization schemes but avoids their weaknesses is of interest. This paper presents a new hybrid optimization algorithm that combines the PSO algorithm and gradient-based local search algorithms to achieve faster convergence and better accuracy of final solution without getting trapped in local minima. In the new gradient-based PSO algorithm, referred to as the GPSO algorithm, the PSO algorithm is used for global exploration and a gradient based scheme is used for accurate local exploration. The global minimum is located by a process of finding progressively better local minima. The GPSO algorithm avoids the use of inertial weights and constriction coefficients which can cause the PSO algorithm to converge to a local minimum if improperly chosen. The De Jong test suite of benchmark optimization problems was used to test the new algorithm and facilitate comparison with the classical PSO algorithm. The GPSO algorithm is compared to four different refinements of the PSO algorithm from the literature and shown to converge faster to a significantly more accurate final solution for a variety of benchmark test functions.