Constructing informative prior distributions from domain knowledge in text classification

  • Authors:
  • Aynur Dayanik;David D. Lewis;David Madigan;Vladimir Menkov;Alexander Genkin

  • Affiliations:
  • Rutgers University, Piscataway, NJ;David D. Lewis Consulting, Chicago, IL;Rutgers University, Piscataway, NJ;Aqsaqal Enterprises, Penticton, B.C. Canada;Rutgers University, Piscataway, NJ

  • Venue:
  • SIGIR '06 Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

Supervised learning approaches to text classification are in practice often required to work with small and unsystematically collected training sets. The alternative to supervised learning is usually viewed to be building classifiers by hand, using a domain expert's understanding of which features of the text are related to the class of interest. This is expensive, requires a degree of sophistication about linguistics and classification, and makes it difficult to use combinations of weak predictors. We propose instead combining domain knowledge with training examples in a Bayesian framework. Domain knowledge is used to specify a prior distribution for the parameters of a logistic regression model, and labeled training data is used to produce a posterior distribution, whose mode we take as the final classifier. We show on three text categorization data sets that this approach can rescue what would otherwise be disastrously bad training situations, producing much more effective classifiers.