Principles of Neural Model Identification, Selection and Adequacy: With Applications in Financial Econometrics
Bayesian learning for neural networks
Bayesian learning for neural networks
A comparison of some error estimates for neural network models
Neural Computation
Confidence interval prediction for neural network models
IEEE Transactions on Neural Networks
Wavelet neural networks: A practical guide
Neural Networks
Hi-index | 0.00 |
Neural networks are a consistent example of non-parametric estimation, with powerful universal approximation properties. However, the effective development and deployment of neural network applications, has to be based on established procedures for estimating confidence and especially prediction intervals. This holds particularly true in cases where there is a strong culture for testing the predictive power of a model, e.g., in financial applications. In this paper we review the major state-of-the-art approaches for constructing confidence and prediction intervals for neural networks, discuss their assumptions, strengths and weaknesses and we compare them in the context of a controlled simulation. Our preliminary results, which are being presented in this paper, indicate a clear superiority of the combination of the bootstrap and maximum likelihood approaches in constructing prediction intervals, relative to the analytical approaches.