Learning to Classify Documents with Only a Small Positive Training Set

  • Authors:
  • Xiao-Li Li;Bing Liu;See-Kiong Ng

  • Affiliations:
  • Institute for Infocomm Research, Heng Mui Keng Terrace, 119613, Singapore;Department of Computer Science, University of Illinois at Chicago, IL 60607-7053,;Institute for Infocomm Research, Heng Mui Keng Terrace, 119613, Singapore

  • Venue:
  • ECML '07 Proceedings of the 18th European conference on Machine Learning
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

Many real-world classification applications fall into the class of positive and unlabeled (PU) learning problems. In many such applications, not only could the negative training examples be missing, the number of positive examples available for learning may also be fairly limited due to the impracticality of hand-labeling a large number of training examples. Current PU learning techniques have focused mostly on identifying reliable negative instances from the unlabeled set U. In this paper, we address the oft-overlooked PU learning problem when the number of training examples in the positive set Pis small. We propose a novel technique LPLP (Learning from Probabilistically Labeled Positive examples) and apply the approach to classify product pages from commercial websites. The experimental results demonstrate that our approach outperforms existing methods significantly, even in the challenging cases where the positive examples in Pand the hidden positive examples in Uwere not drawn from the same distribution.