2005 Special Issue: Bayesian approach to feature selection and parameter tuning for support vector machine classifiers

  • Authors:
  • Carl Gold;Alex Holub;Peter Sollich

  • Affiliations:
  • Computation and Neural Systems, California Institute of Technology, 139-74, Pasadena, CA 91125, USA;Computation and Neural Systems, California Institute of Technology, 139-74, Pasadena, CA 91125, USA;Department of Mathematics, King's College London, Strand, London WC2R 2LS, UK

  • Venue:
  • Neural Networks - 2005 Special issue: IJCNN 2005
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

A Bayesian point of view of SVM classifiers allows the definition of a quantity analogous to the evidence in probabilistic models. By maximizing this one can systematically tune hyperparameters and, via automatic relevance determination (ARD), select relevant input features. Evidence gradients are expressed as averages over the associated posterior and can be approximated using Hybrid Monte Carlo (HMC) sampling. We describe how a Nystrom approximation of the Gram matrix can be used to speed up sampling times significantly while maintaining almost unchanged classification accuracy. In experiments on classification problems with a significant number of irrelevant features this approach to ARD can give a significant improvement in classification performance over more traditional, non-ARD, SVM systems. The final tuned hyperparameter values provide a useful criterion for pruning irrelevant features, and we define a measure of relevance with which to determine systematically how many features should be removed. This use of ARD for hard feature selection can improve classification accuracy in non-ARD SVMs. In the majority of cases, however, we find that in data sets constructed by human domain experts the performance of non-ARD SVMs is largely insensitive to the presence of some less relevant features. Eliminating such features via ARD then does not improve classification accuracy, but leads to impressive reductions in the number of features required, by up to 75%.