Language models as representations for weakly-supervised NLP tasks

  • Authors:
  • Fei Huang;Alexander Yates;Arun Ahuja;Doug Downey

  • Affiliations:
  • Temple University, Philadelphia, PA;Temple University, Philadelphia, PA;Northwestern University, Evanston, IL;Northwestern University, Evanston, IL

  • Venue:
  • CoNLL '11 Proceedings of the Fifteenth Conference on Computational Natural Language Learning
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Finding the right representation for words is critical for building accurate NLP systems when domain-specific labeled data for the task is scarce. This paper investigates language model representations, in which language models trained on unlabeled corpora are used to generate real-valued feature vectors for words. We investigate ngram models and probabilistic graphical models, including a novel lattice-structured Markov Random Field. Experiments indicate that language model representations outperform traditional representations, and that graphical model representations outperform ngram models, especially on sparse and polysemous words.