Efficient computations for large least square support vector machine classifiers

  • Authors:
  • Kok Seng Chua

  • Affiliations:
  • Data Analysis and Data Intensive Computing Group, High End Computing Programme, Institute of High Performance Computing, 1 Science Park Road, #01-01 The Capricorn, Singapore Science Park II, Singa ...

  • Venue:
  • Pattern Recognition Letters
  • Year:
  • 2003

Quantified Score

Hi-index 0.10

Visualization

Abstract

We observed that the linear system in the training of the least square support vector machine (LSSVM) proposed by Suykens and Vandewalle (Neural process. Lett. 9 (1999a) 293-300; IEEE Trans. Neural Networks 10(4) (1999b) 907-912) can be placed in a more symmetric form so that for a data set with N data points and m features, the linear system can be solved by inverting an m × m instead of an N × N matrix and storing and working with matrices of size at most m × N. This allows us to apply LSSVM to very large data set with small number of features. Our computations show that a data set with a million data points and 10 features can be trained in only 45 s. We also compared the effectiveness and efficiency of our method to standard LSSVM and standard SVM. An example using a quadratic kernel is also given.