SVM paradoxes

  • Authors:
  • Jean Beney;Cornelis H. A. Koster

  • Affiliations:
  • LCI, Département Informatique, INSA de Lyon F69621 Villeurbanne, Université de Lyon;ICIS, Radboud University, Nijmegen

  • Venue:
  • PSI'09 Proceedings of the 7th international Andrei Ershov Memorial conference on Perspectives of Systems Informatics
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Support Vector Machines (SVM) is widely considered to be the best algorithm for text classification because it is based on a well-founded theory (SRM): in the separable case it provides the best result possible for a given set of separation functions, and therefore it does not require tuning. In this paper we scrutinize these suppositions, and encounter some paradoxes. In a large-scale experiment it is shown that even in the separable case SVM's extension to non-separable data may give a better result by minimizing the confidence interval of the risk. However, the use of this extension necessitates the tuning of the complexity constant. Furthermore, the use of SVM for optimizing precision and recall through the F function necessitates the tuning of the threshold found by SVM. But the tuned classifier does not generalize well. Furthermore, a more precise definition is given to the notion of training errors.