Several SVM Ensemble Methods Integrated with Under-Sampling for Imbalanced Data Learning

  • Authors:
  • Zhiyong Lin;Zhifeng Hao;Xiaowei Yang;Xiaolan Liu

  • Affiliations:
  • School of Computer Science and Engineering, South China University of Technology, Guangzhou 510640 and Department of Computer Science, Guangdong Polytechnic Normal University, Guangzhou 510665;Guangdong University of Technology, Guangzhou 510006;College of Science, South China University of Technology, Email: sophyca@yahoo.cn, Guangzhou, 510640;College of Science, South China University of Technology, Email: sophyca@yahoo.cn, Guangzhou, 510640

  • Venue:
  • ADMA '09 Proceedings of the 5th International Conference on Advanced Data Mining and Applications
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Imbalanced data learning (IDL) is one of the most active and important fields in machine learning research. This paper focuses on exploring the efficiencies of four different SVM ensemble methods integrated with under-sampling in IDL. The experimental results on 20 UCI imbalanced datasets show that two new ensemble algorithms proposed in this paper, i.e., CABagE (which is bagging-style) and MABstE (which is boosting-style), can output the SVM ensemble classifiers with better minority-class-recognition abilities than the existing ensemble methods. Further analysis on the experimental results indicates that MABstE has the best overall classification performance, and we believe that this should be attributed to its more robust example-weighting mechanism.