A framework for evaluating approximation methods for Gaussian process regression

  • Authors:
  • Krzysztof Chalupka;Christopher K. I. Williams;Iain Murray

  • Affiliations:
  • Computation and Neural Systems, California Institute of Technology, Pasadena, CA;School of Informatics, University of Edinburgh, Edinburgh, UK;School of Informatics, University of Edinburgh, Edinburgh, UK

  • Venue:
  • The Journal of Machine Learning Research
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Gaussian process (GP) predictors are an important component of many Bayesian approaches to machine learning. However, even a straightforward implementation of Gaussian process regression (GPR) requires O(n2) space and O(n3) time for a data set of n examples. Several approximation methods have been proposed, but there is a lack of understanding of the relative merits of the different approximations, and in what situations they are most useful. We recommend assessing the quality of the predictions obtained as a function of the compute time taken, and comparing to standard baselines (e.g., Subset of Data and FITC). We empirically investigate four different approximation algorithms on four different prediction problems, and make our code available to encourage future comparisons.