An empirical evaluation of bagging with different algorithms on imbalanced data

  • Authors:
  • Guohua Liang;Chengqi Zhang

  • Affiliations:
  • The Centre for Quantum Computation & Intelligent Systems, FEIT, University of Technology, Sydney, NSW, Australia;The Centre for Quantum Computation & Intelligent Systems, FEIT, University of Technology, Sydney, NSW, Australia

  • Venue:
  • ADMA'11 Proceedings of the 7th international conference on Advanced Data Mining and Applications - Volume Part I
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

This study investigates the effectiveness of bagging with respect to different learning algorithms on Imbalanced data-sets. The purpose of this research is to investigate the performance of bagging based on two unique approaches: (1) classify base learners with respect to 12 different learning algorithms in general terms, and (2) evaluate the performance of bagging predictors on data with imbalanced class distributions. The former approach develops a method to categorize base learners by using two-dimensional robustness and stability decomposition on 48 benchmark data-sets; while the latter approach investigates the performance of bagging predictors by using evaluation metrics, True Positive Rate (TPR ), Geometric mean (G-mean ) for the accuracy on the majority and minority classes, and the Receiver Operating Characteristic (ROC ) curve on 12 imbalanced data-sets. Our studies assert that both stability and robustness are important factors for building high performance bagging predictors on data with imbalanced class distributions. The experimental results demonstrated that PART and Multi-layer Proceptron (MLP) are the learning algorithms with the best bagging performance on 12 imbalanced data-sets. Moreover, only four out of 12 bagging predictors are statistically superior to single learners based on both G-mean and TPR evaluation metrics over 12 imbalanced data-sets.