Building high-performance classifiers using positive and unlabeled examples for text classification

  • Authors:
  • Ting Ke;Bing Yang;Ling Zhen;Junyan Tan;Yi Li;Ling Jing

  • Affiliations:
  • Department of Applied Mathematics, College of Science, China Agricultural University, Beijing, P.R. China;Department of Applied Mathematics, College of Science, China Agricultural University, Beijing, P.R. China;Department of Applied Mathematics, College of Science, China Agricultural University, Beijing, P.R. China;Department of Applied Mathematics, College of Science, China Agricultural University, Beijing, P.R. China;Department of Mathematics, School of Science, Beijing University of Posts and Telecommunications, Beijing, P.R. China;Department of Applied Mathematics, College of Science, China Agricultural University, Beijing, P.R. China

  • Venue:
  • ISNN'12 Proceedings of the 9th international conference on Advances in Neural Networks - Volume Part II
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper studies the problem of building text classifiers using only positive and unlabeled examples. At present, many techniques for solving this problem were proposed, such as Biased-SVM which is the existing popular method and its classification performance is better than most of two-step techniques. In this paper, an improved iterative classification approach is proposed which is the extension of Biased-SVM. The first iteration of our developed approach is Biased-SVM and the next iterations are to identify confident positive examples from the unlabeled examples. Then an extra penalty factor is given to weight these confident positive examples error. Experiments show that it is effective for text classification and outperforms the Biased-SVM and other two step techniques.