An Evolutionary Multi-objective Neural Network Optimizer with Bias-Based Pruning Heuristic

  • Authors:
  • Prospero C. Naval, Jr.;John Paul Yusiong

  • Affiliations:
  • Department of Computer Science, University of the Philippines-Diliman, Diliman, Quezon City, Philippines;Division of Natural Sciences and Mathematics, University of the, Philippines-Visayas, Tacloban City, Leyte, Philippines

  • Venue:
  • ISNN '07 Proceedings of the 4th international symposium on Neural Networks: Advances in Neural Networks, Part III
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

Neural network design aims for high classification accuracy and low network architecture complexity. It is also known that simultaneous optimization of both model accuracy and complexity improves generalization while avoiding overfitting on data. We describe a neural network training procedure that uses multi-objective optimization to evolve networks which are optimal both with respect to classification accuracy and architectural complexity. The NSGA-II algorithm is employed to evolve a population of neural networks that are minimal in both training error and a Minimum Description Length-based network complexity measure. We further propose a pruning rule based on the following heuristic: connections to or from a node may be severed if their weight values are smaller than the network's smallest bias. Experiments on benchmark datasets show that the proposed evolutionary multi-objective approach to neural network design employing the bias-based pruning heuristic yields networks that have far fewer connections without seriously compromising generalization performance when compared to other existing evolutionary optimization algorithms.