A new constructive algorithm for architectural and functional adaptation of artificial neural networks

  • Authors:
  • Md. Monirul Islam;Md. Abdus Sattar;Md. Faijul Amin;Xin Yao;Kazuyuki Murase

  • Affiliations:
  • Department of Computer Science and Engineering, Bangladesh University of Engineering and Technology, Dhaka, Bangladesh and Department of Human and Artificial Intelligence Systems, Graduate School ...;Department of Computer Science and Engineering, Bangladesh University of Engineering and Technology, Dhaka, Bangladesh;Department of Human and Artificial Intelligence Systems, University of Fukui, Fukui, Japan;Center of Excellence for Research in Computational Intelligence and Applications, School of Computer Science, University of Birmingham, Birmingham, UK;Department of Human and Artificial Intelligence Systems, Graduate School of Engineering, University of Fukui, Fukui, Japan and Research and Education Program for Life Science, University of Fukui, ...

  • Venue:
  • IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

The generalization ability of artificial neural networks (ANNs) is greatly dependent on their architectures. Constructive algorithms provide an attractive automatic way of determining a near-optimal ANN architecture for a given problem. Several such algorithms have been proposed in the literature and shown their effectiveness. This paper presents a new constructive algorithm (NCA) in automatically determining ANN architectures. Unlike most previous studies on determining ANN architectures, NCA puts emphasis on architectural adaptation and functional adaptation in its architecture determination process. It uses a constructive approach to determine the number of hidden layers in an ANN and of neurons in each hidden layer. To achieve functional adaptation, NCA trains hidden neurons in the ANN by using different training sets that were created by employing a similar concept used in the boosting algorithm. The purpose of using different training sets is to encourage hidden neurons to learn different parts or aspects of the training data so that the ANN can learn the whole training data in a better way. In this paper, the convergence and computational issues of NCA are analytically studied. The computational complexity of NCA is found to be O(W × Pt × τ), where W is the number of weights in the ANN, Pt is the number of training examples, and τ is the number of training epochs. This complexity has the same order as what the backpropagation learning algorithm requires for training a fixed ANN architecture. A set of eight classification and two approximation benchmark problems was used to evaluate the performance of NCA. The experimental results show that NCA can produce ANN architectures with fewer hidden neurons and better generalization ability compared to existing constructive and nonconstructive algorithms.