Exhaustive learning

  • Authors:
  • D. B. Schwartz;V. K. Samalam;Sara A. Solla;J. S. Denker

  • Affiliations:
  • GTE Laboratories, Waltham, MA 02254 USA;GTE Laboratories, Waltham, MA 02254 USA;AT&T Bell Laboratories, Holmdel, NJ 07733 USA;AT&T Bell Laboratories, Holmdel, NJ 07733 USA

  • Venue:
  • Neural Computation
  • Year:
  • 1990

Quantified Score

Hi-index 0.00

Visualization

Abstract

Exhaustive exploration of an ensemble of networks is used to model learning and generalization in layered neural networks. A simple Boolean learning problem involving networks with binary weights is numerically solved to obtain the entropy Sm and the average generalization ability Gm as a function of the size m of the training set. Learning curves Gm vs m are shown to depend solely on the distribution of generalization abilities over the ensemble of networks. Such distribution is determined prior to learning, and provides a novel theoretical tool for the prediction of network performance on a specific task.