Machine Learning
Neural Networks for Pattern Recognition
Neural Networks for Pattern Recognition
Gaussian Processes for Machine Learning (Adaptive Computation and Machine Learning)
Gaussian Processes for Machine Learning (Adaptive Computation and Machine Learning)
A Unifying View of Sparse Approximate Gaussian Process Regression
The Journal of Machine Learning Research
Bagging for Gaussian process regression
Neurocomputing
Some new results on neural network approximation
Neural Networks
Applying falsity input to neural networks to solve single output regression problems
ACACOS'11 Proceedings of the 10th WSEAS international conference on Applied computer and applied computational science
Hi-index | 0.00 |
For regression tasks, traditional neural networks (NNs) have been superseded by Gaussian processes, which provide probabilistic predictions (input-dependent error bars), improved accuracy, and virtually no overfitting. Due to their high computational cost, in scenarios with massive data sets, one has to resort to sparse Gaussian processes, which strive to achieve similar performance with much smaller computational effort. In this context, we introduce a mixture of NNs with marginalized output weights that can both provide probabilistic predictions and improve on the performance of sparse Gaussian processes, at the same computational cost. The effectiveness of this approach is shown experimentally on some representative large data sets.