Universal approximation using radial-basis-function networks
Neural Computation
Approximation and radial-basis-function networks
Neural Computation
Approximation by superposition of sigmoidal and radial basis functions
Advances in Applied Mathematics
Tractability and strong tractability of linear multivariate problems
Journal of Complexity
Dimension-independent bounds on the degree of approximation by neural networks
IBM Journal of Research and Development
Complexity and information
On best approximation by ridge functions
Journal of Approximation Theory
Error bounds for approximation with neural networks
Journal of Approximation Theory
Dynamic Programming
On the tractability of multivariate integration and approximation by neural networks
Journal of Complexity
Minimization of error functionals over perceptron networks
Neural Computation
An integral upper bound for neural network approximation
Neural Computation
Comparison of worst case errors in linear and neural network approximation
IEEE Transactions on Information Theory
Universal approximation bounds for superpositions of a sigmoidal function
IEEE Transactions on Information Theory
Model Complexity of Neural Networks and Integral Transforms
ICANN '09 Proceedings of the 19th International Conference on Artificial Neural Networks: Part I
On tractability of neural-network approximation
ICANNGA'09 Proceedings of the 9th international conference on Adaptive and natural computing algorithms
Some comparisons of model complexity in linear and neural-network approximation
ICANN'10 Proceedings of the 20th international conference on Artificial neural networks: Part III
Kernel networks with fixed and variable widths
ICANNGA'11 Proceedings of the 10th international conference on Adaptive and natural computing algorithms - Volume Part I
Bounds for approximate solutions of Fredholm integral equations using kernel networks
ICANN'11 Proceedings of the 21th international conference on Artificial neural networks - Volume Part I
Error estimates of quasi-interpolation and its derivatives
Journal of Computational and Applied Mathematics
Some comparisons of networks with radial and kernel units
ICANN'12 Proceedings of the 22nd international conference on Artificial Neural Networks and Machine Learning - Volume Part II
Hi-index | 0.00 |
Complexity of Gaussian-radial-basis-function networks, with varying widths, is investigated. Upper bounds on rates of decrease of approximation errors with increasing number of hidden units are derived. Bounds are in terms of norms measuring smoothness (Bessel and Sobolev norms) multiplied by explicitly given functions a(r,d) of the number of variables d and degree of smoothness r. Estimates are proven using suitable integral representations in the form of networks with continua of hidden units computing scaled Gaussians and translated Bessel potentials. Consequences on tractability of approximation by Gaussian-radial-basis function networks are discussed.