Efficient estimation of neural weights by polynomial approximation

  • Authors:
  • G. Ritter

  • Affiliations:
  • Fakultat fur Math. und Inf., Passau Univ.

  • Venue:
  • IEEE Transactions on Information Theory
  • Year:
  • 2006

Quantified Score

Hi-index 754.84

Visualization

Abstract

It has been known for some years that the uniform-density problem for forward neural networks has a positive answer: any real-valued, continuous function on a compact subset of Rd can be uniformly approximated by a sigmoidal neural network with one hidden layer. We design here algorithms for efficient uniform approximation by a certain class of neural networks with one hidden layer which we call nearly exponential. This class contains, e.g., all networks with the activation functions 1/(1+e-t), tanh(t), or et ∧1 in their hidden layers. The algorithms flow from a theorem stating that such networks attain the order of approximation O(N-1 d/), d being dimension and N the number of hidden neurons. This theorem, in turn, is a consequence of a close relationship between neural networks of nearly exponential type and multivariate algebraic and exponential polynomials. The algorithms need neither a starting point nor learning parameters; they do not get stuck in local minima, and the gain in execution time relative to the backpropagation algorithm is enormous. The size of the hidden layer can be bounded analytically as a function of the precision required