Computational properties of probabilistic neural-networks

  • Authors:
  • Jiří Grim;Jan Hora

  • Affiliations:
  • Institute of Information Theory and Automation, Prague 8, Czech Republic;Institute of Information Theory and Automation, Prague 8, Czech Republic

  • Venue:
  • ICANN'10 Proceedings of the 20th international conference on Artificial neural networks: Part III
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

We discuss the problem of overfitting of probabilistic neural networks in the framework of statistical pattern recognition. The probabilistic approach to neural networks provides a statistically justified subspace method of classification. The underlying structural mixture model includes binary structural parameters and can be optimized by EM algorithm in full generality. Formally, the structural model reduces the number of parameters included and therefore the structural mixtures become less complex and less prone to overfitting. We illustrate how recognition accuracy and the effect of overfitting is influenced by mixture complexity and by the size of training data set.