On large scale nonlinear least squares calculations
SIAM Journal on Scientific and Statistical Computing
Separable nonlinear least squares with multiple right-hand sides
SIAM Journal on Matrix Analysis and Applications
Fast exact multiplication by the Hessian
Neural Computation
Advanced algorithms for neural networks: a C++ sourcebook
Advanced algorithms for neural networks: a C++ sourcebook
Applied numerical linear algebra
Applied numerical linear algebra
Natural gradient works efficiently in learning
Neural Computation
An Adaptive Nonlinear Least-Squares Algorithm
ACM Transactions on Mathematical Software (TOMS)
Trust-region methods
IJCNN '00 Proceedings of the IEEE-INNS-ENNS International Joint Conference on Neural Networks (IJCNN'00)-Volume 2 - Volume 2
Proceedings of the CUBE International Information Technology Conference
Hi-index | 0.00 |
This paper briefly introduces our numerical linear algebra approaches for solving structured nonlinear least squares problems arising from 'multiple-output' neural-network (NN) models. Our algorithms feature trust-region regularization, and exploit sparsity of either the 'blockangular' residual Jacobian matrix or the 'block-arrow' Gauss-Newton Hessian (or Fisher information matrix in statistical sense) depending on problem scale so as to render a large class of NN-learning algorithms 'efficient' in both memory and operation costs. Using a relatively large real-world nonlinear regression application, we shall explain algorithmic strengths and weaknesses, analyzing simulation results obtained by both direct and iterative trust-region algorithms with two distinct NN models: 'multilayer perceptrons' (MLP) and 'complementary mixtures of MLP-experts' (or neuro-fuzzy modular networks).