Minimum Class Variance Support Vector Machines

  • Authors:
  • S. Zafeiriou;A. Tefas;I. Pitas

  • Affiliations:
  • Aristotle Univ. of Thessaloniki, Thessaloniki;-;-

  • Venue:
  • IEEE Transactions on Image Processing
  • Year:
  • 2007

Quantified Score

Hi-index 0.01

Visualization

Abstract

In this paper, a modified class of support vector machines (SVMs) inspired from the optimization of Fisher's discriminant ratio is presented, the so-called minimum class variance SVMs (MCVSVMs). The MCVSVMs optimization problem is solved in cases in which the training set contains less samples that the dimensionality of the training vectors using dimensionality reduction through principal component analysis (PCA). Afterward, the MCVSVMs are extended in order to find nonlinear decision surfaces by solving the optimization problem in arbitrary Hilbert spaces defined by Mercer's kernels. In that case, it is shown that, under kernel PCA, the nonlinear optimization problem is transformed into an equivalent linear MCVSVMs problem. The effectiveness of the proposed approach is demonstrated by comparing it with the standard SVMs and other classifiers, like kernel Fisher discriminant analysis in facial image characterization problems like gender determination, eyeglass, and neutral facial expression detection.