Enhance Neural Networks Training Using GA with Chaos Theory

  • Authors:
  • K. Y. Leong;Augustina Sitiol;Kalaiarasi Sonai Anbananthen

  • Affiliations:
  • School of Engineering and Information Technology, University Malaysia Sabah, Kota Kinabalu, Malaysia 88999;School of Engineering and Information Technology, University Malaysia Sabah, Kota Kinabalu, Malaysia 88999;School of Engineering and Information Technology, University Malaysia Sabah, Kota Kinabalu, Malaysia 88999

  • Venue:
  • ISNN 2009 Proceedings of the 6th International Symposium on Neural Networks: Advances in Neural Networks - Part II
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

There are numerous algorithms available for training artificial neural networks. Besides classical algorithms for supervised learning such as backpropagation, associative memory and radial basis function, this training task can be employed by evolutionary computation since most of the gradient descent related algorithms can be view as an application of optimization theory and stochastic search. In this paper, the logistic model of population growth from ecology is integrated into initialization, selection and crossover operators of genetic algorithms for neural network training. These chaotic operators are very efficient in maintaining the population diversity during the evolution process of genetic algorithms. A comparison is done on the basis of a benchmark comprising several data classification problems for neural networks. Three variants of training --- Backpropagation (BP), Genetic Algorithms (GA) and Genetic Algorithms with Chaotic Operators (GACO) --- are described and compared. The experimental results confirm the dynamic mobility of chaotic algorithms in GACO network training, which can overcome saturation and improve the convergence rate.