Large Scale Kernel Regression via Linear Programming

  • Authors:
  • O. L. Mangasarian;David R. Musicant

  • Affiliations:
  • Computer Sciences Department, University of Wisconsin, 1210 West Dayton Street, Madison, WI 53706, USA. olvi@cs.wisc.edu;Department of Mathematics and Computer Science, Carleton College, One North College Street, Northfield, MN 55057, USA. dmusican@carleton.edu

  • Venue:
  • Machine Learning
  • Year:
  • 2002

Quantified Score

Hi-index 0.00

Visualization

Abstract

The problem of tolerant data fitting by a nonlinear surface, induced by a kernel-based support vector machine is formulated as a linear program with fewer number of variables than that of other linear programming formulations. A generalization of the linear programming chunking algorithm for arbitrary kernels is implemented for solving problems with very large datasets wherein chunking is performed on both data points and problem variables. The proposed approach tolerates a small error, which is adjusted parametrically, while fitting the given data. This leads to improved fitting of noisy data (over ordinary least error solutions) as demonstrated computationally. Comparative numerical results indicate an average time reduction as high as 26.0% over other formulations, with a maximal time reduction of 79.7%. Additionally, linear programs with as many as 16,000 data points and more than a billion nonzero matrix elements are solved.