MSMOTE: Improving Classification Performance When Training Data is Imbalanced

  • Authors:
  • Shengguo Hu;Yanfeng Liang;Lintao Ma;Ying He

  • Affiliations:
  • -;-;-;-

  • Venue:
  • IWCSE '09 Proceedings of the 2009 Second International Workshop on Computer Science and Engineering - Volume 02
  • Year:
  • 2009

Quantified Score

Hi-index 0.01

Visualization

Abstract

Learning from data sets that contain very few instances of the minority class usually produces biased classifiers that have a higher predictive accuracy over the majority class, but poorer predictive accuracy over the minority class. SMOTE (Synthetic Minority Over-sampling Technique) is specifically designed for learning from imbalanced data sets. This paper presents a modified approach (MSMOTE) for learning from imbalanced data sets, based on the SMOTE algorithm. MSMOTE not only considers the distribution of minority class samples, but also eliminates noise samples by adaptive mediation. The combination of MSMOTE and AdaBoost are applied to several highly and moderately imbalanced data sets. The experimental results show that the prediction performance of MSMOTE is better than SMOTEBoost in the minority class and F-values are also improved.