Overcoming the local-minimum problem in training multilayer perceptrons with the NRAE training method

  • Authors:
  • James Ting-Ho Lo;Yichuan Gui;Yun Peng

  • Affiliations:
  • Department of Mathematics and Statistics, University of Maryland, Baltimore County, Baltimore, Maryland;Department of Computer Science and Electrical Engineering, University of Maryland, Baltimore County, Baltimore, Maryland;Department of Computer Science and Electrical Engineering, University of Maryland, Baltimore County, Baltimore, Maryland

  • Venue:
  • ISNN'12 Proceedings of the 9th international conference on Advances in Neural Networks - Volume Part I
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

A method of training multilayer perceptrons (MLPs) to reach a global or nearly global minimum of the standard mean squared error (MSE) criterion is proposed. It has been found that the region in the weight space that does not have a local minimum of the normalized risk-averting error (NRAE) criterion expands strictly to the entire weight space as the risk-sensitivity index increases to infinity. If the MLP under training has enough hidden neurons, the MSE and NRAE criteria are both equal to nearly zero at a global or nearly global minimum. Training the MLP with the NRAE at a sufficiently large risk-sensitivity index can therefore effectively avoid non-global local minima. Numerical experiments show consistently successful convergence from different initial guesses of the weights of the MLP at a risk-sensitivity index over 106. The experiments are conducted on examples with non-global local minima of the MSE criterion that are difficult to escape from by training directly with the MSE criterion.