Modeling Documents by Combining Semantic Concepts with Unsupervised Statistical Learning

  • Authors:
  • Chaitanya Chemudugunta;America Holloway;Padhraic Smyth;Mark Steyvers

  • Affiliations:
  • Department of Computer Science, University of California,Irvine, Irvine,;Department of Computer Science, University of California,Irvine, Irvine,;Department of Computer Science, University of California,Irvine, Irvine,;Department of Cognitive Science, University of California, Irvine, Irvine,

  • Venue:
  • ISWC '08 Proceedings of the 7th International Conference on The Semantic Web
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

Human-defined concepts are fundamental building-blocks in constructing knowledge bases such as ontologies. Statistical learning techniques provide an alternative automated approach to concept definition, driven by data rather than prior knowledge. In this paper we propose a probabilistic modeling framework that combines both human-defined concepts and data-driven topics in a principled manner. The methodology we propose is based on applications of statistical topic models (also known as latent Dirichlet allocation models). We demonstrate the utility of this general framework in two ways. We first illustrate how the methodology can be used to automatically tag Web pages with concepts from a known set of concepts without any need for labeled documents. We then perform a series of experiments that quantify how combining human-defined semantic knowledge with data-driven techniques leads to better language models than can be obtained with either alone.