Exploiting probabilistic topic models to improve text categorization under class imbalance

  • Authors:
  • Enhong Chen;Yanggang Lin;Hui Xiong;Qiming Luo;Haiping Ma

  • Affiliations:
  • School of Computer Science and Technology, P.O. Box 4, Hefei, Anhui 230027, PR China;School of Computer Science and Technology, P.O. Box 4, Hefei, Anhui 230027, PR China;Department of Management Science and Information Systems, Rutgers, The State University of New Jersey, New Brunswick, NJ 08901-8554, USA;School of Computer Science and Technology, P.O. Box 4, Hefei, Anhui 230027, PR China;School of Computer Science and Technology, P.O. Box 4, Hefei, Anhui 230027, PR China

  • Venue:
  • Information Processing and Management: an International Journal
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

In text categorization, it is quite often that the numbers of documents in different categories are different, i.e., the class distribution is imbalanced. We propose a unique approach to improve text categorization under class imbalance by exploiting the semantic context in text documents. Specifically, we generate new samples of rare classes (categories with relatively small amount of training data) by using global semantic information of classes represented by probabilistic topic models. In this way, the numbers of samples in different categories can become more balanced and the performance of text categorization can be improved using this transformed data set. Indeed, the proposed method is different from traditional re-sampling methods, which try to balance the number of documents in different classes by re-sampling the documents in rare classes. Such re-sampling methods can cause overfitting. Another benefit of our approach is the effective handling of noisy samples. Since all the new samples are generated by topic models, the impact of noisy samples is dramatically reduced. Finally, as demonstrated by the experimental results, the proposed methods can achieve better performance under class imbalance and is more tolerant to noisy samples.