Efficient Block Training of Multilayer Perceptrons

  • Authors:
  • A. Navia-Vázquez;A. R. Figueiras-vidal

  • Affiliations:
  • DTC, Universidad Carlos III de Madrid, 28911-Leganés (Madrid), Spain;DTC, Universidad Carlos III de Madrid, 28911-Leganés (Madrid), Spain

  • Venue:
  • Neural Computation
  • Year:
  • 2000

Quantified Score

Hi-index 0.00

Visualization

Abstract

The attractive possibility of applying layerwise block training algorithms to multilayer perceptrons MLP, which offers initial advantages in computational effort, is refined in this article by means of introducing a sensitivity correction factor in the formulation. This results in a clear performance advantage, which we verify in several applications. The reasons for this advantage are discussed and related to implicit relations with second-order techniques, natural gradient formulations through Fisher's information matrix, and sample selection. Extensions to recurrent networks and other research lines are suggested at the close of the article.