Extreme re-balancing for SVMs: a case study

  • Authors:
  • Bhavani Raskutti;Adam Kowalczyk

  • Affiliations:
  • Telstra Corporation, Clayton, Victoria, Australia;Telstra Corporation, Clayton, Victoria, Australia

  • Venue:
  • ACM SIGKDD Explorations Newsletter - Special issue on learning from imbalanced datasets
  • Year:
  • 2004

Quantified Score

Hi-index 0.01

Visualization

Abstract

There are many practical applications where learning from single class examples is either, the only possible solution, or has a distinct performance advantage. The first case occurs when obtaining examples of a second class is difficult, e.g., classifying sites of "interest" based on web accesses. The second situation is exemplified by the gene knock-out experiments for understanding Aryl Hydrocarbon Receptor signalling pathway that provided the data for the second task of the KDD 2002 Cup, where minority one-class SVMs significantly outperform models learnt using examples from both classes.This paper explores the limits of supervised learning of a two class discrimination from data with heavily unbalanced class proportions. We focus on the case of supervised learning with support vector machines. We consider the impact of both sampling and weighting imbalance compensation techniques and then extend the balancing to extreme situations when one of the classes is ignored completely and the learning is accomplished using examples from a single class.Our investigation with the data for KDD 2002 Cup as well as text benchmarks such as Reuters Newswire shows that there is a consistent pattern of performance differences between one and two-class learning for all SVMs investigated, and these patterns persist even with aggressive dimensionality reduction through automated feature selection. Using insight gained from the above analysis, we generate synthetic data showing similar pattern of performance.