Backpropagation uses prior information efficiently

  • Authors:
  • E. Barnard;E. C. Botha

  • Affiliations:
  • Dept. of Comput. Sci. & Eng., Oregon Graduate Inst., Portland, OR;-

  • Venue:
  • IEEE Transactions on Neural Networks
  • Year:
  • 1993

Quantified Score

Hi-index 0.00

Visualization

Abstract

The ability of neural net classifiers to deal with a priori information is investigated. For this purpose, backpropagation classifiers are trained with data from known distributions with variable a priori probabilities, and their performance on separate test sets is evaluated. It is found that backpropagation employs a priori information in a slightly suboptimal fashion, but this does not have serious consequences on the performance of the classifier. Furthermore, it is found that the inferior generalization that results when an excessive number of network parameters are used can (partially) be ascribed to this suboptimality