2005 Special Issue: Constructing Bayesian formulations of sparse kernel learning methods

  • Authors:
  • Gavin C. Cawley;Nicola L. C. Talbot

  • Affiliations:
  • School of Computing Sciences, University of East Anglia, Norwich NR4 7TJ, UK;School of Computing Sciences, University of East Anglia, Norwich NR4 7TJ, UK

  • Venue:
  • Neural Networks - 2005 Special issue: IJCNN 2005
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

We present here a simple technique that simplifies the construction of Bayesian treatments of a variety of sparse kernel learning algorithms. An incomplete Cholesky factorisation is employed to modify the dual parameter space, such that the Gaussian prior over the dual model parameters is whitened. The regularisation term then corresponds to the usual weight-decay regulariser, allowing the Bayesian analysis to proceed via the evidence framework of MacKay. There is in addition a useful by-product associated with the incomplete Cholesky factorisation algorithm, it also identifies a subset of the training data forming an approximate basis for the entire dataset in the kernel-induced feature space, resulting in a sparse model. Bayesian treatments of the kernel ridge regression (KRR) algorithm, with both constant and heteroscedastic (input dependent) variance structures, and kernel logistic regression (KLR) are provided as illustrative examples of the proposed method, which we hope will be more widely applicable.