Training neural networks for classification using growth probability-based evolution

  • Authors:
  • J. H. Ang;K. C. Tan;A. Al-Mamun

  • Affiliations:
  • Department of Electrical and Computer Engineering, National University of Singapore, 4 Engineering Drive 3, Singapore 117576, Singapore;Department of Electrical and Computer Engineering, National University of Singapore, 4 Engineering Drive 3, Singapore 117576, Singapore;Department of Electrical and Computer Engineering, National University of Singapore, 4 Engineering Drive 3, Singapore 117576, Singapore

  • Venue:
  • Neurocomputing
  • Year:
  • 2008

Quantified Score

Hi-index 0.01

Visualization

Abstract

In this paper, a novel evolutionary algorithm (EA) based on a newly formulated parameter, i.e., growth probability (P"g) is used to evolve the near optimal weights and the number of hidden neurons in neural networks (NNs). Training NNs with growth probability based evolution (NN-GP) initializes networks with only one hidden neuron and the networks are allowed to grow until a suitable size. Growing of neurons is not restricted to one hidden neuron at a time as the optimal number of hidden neurons for the NNs might be a few neurons more than what it represents now. If this solution in the search space is far, networks have to add several number of hidden neurons. Growth rate is based on Gaussian distribution thus providing a way to escape local optima. A self-adaptive version (NN-SAGP) with the aim of evolving the growth probability in parallel with NNs during each generation is also proposed. The evolved networks are applied to widely used real-world benchmark problems. Simulation results show that the proposed approach is effective for evolving NNs with good classification accuracy and low complexity.