Hierarchical linear support vector machine

  • Authors:
  • Irene Rodriguez-Lujan;Carlos Santa Cruz;Ramon Huerta

  • Affiliations:
  • Departamento de Ingenierí/a Informá/tica and Instituto de Ingenierí/a del Conocimiento, C/Francisco Tomá/s y Valiente, 11/ E.P.S., Edificio B/ UAM-Cantoblanco, 28049 Madrid, Spain;Departamento de Ingenierí/a Informá/tica and Instituto de Ingenierí/a del Conocimiento, C/Francisco Tomá/s y Valiente, 11/ E.P.S., Edificio B/ UAM-Cantoblanco, 28049 Madrid, Spain;BioCircuits Institute, University of California San Diego, La Jolla, CA 92093-0404, USA

  • Venue:
  • Pattern Recognition
  • Year:
  • 2012

Quantified Score

Hi-index 0.01

Visualization

Abstract

The increasing size and dimensionality of real-world datasets make it necessary to design efficient algorithms not only in the training process but also in the prediction phase. In applications such as credit card fraud detection, the classifier needs to predict an event in 10ms at most. In these environments the speed of the prediction constraints heavily outweighs the training costs. We propose a new classification method, called a Hierarchical Linear Support Vector Machine (H-LSVM), based on the construction of an oblique decision tree in which the node split is obtained as a Linear Support Vector Machine. Although other methods have been proposed to break the data space down in subregions to speed up Support Vector Machines, the H-LSVM algorithm represents a very simple and efficient model in training but mainly in prediction for large-scale datasets. Only a few hyperplanes need to be evaluated in the prediction step, no kernel computation is required and the tree structure makes parallelization possible. In experiments with medium and large datasets, the H-LSVM reduces the prediction cost considerably while achieving classification results closer to the non-linear SVM than that of the linear case.