Semi-supervised sequence modeling with syntactic topic models

  • Authors:
  • Wei Li;Andrew McCallum

  • Affiliations:
  • Computer Science Department, University of Massachusetts, Amherst;Computer Science Department, University of Massachusetts, Amherst

  • Venue:
  • AAAI'05 Proceedings of the 20th national conference on Artificial intelligence - Volume 2
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

Although there has been significant previous work on semi-supervised learning for classification, there has been relatively little in sequence modeling. This paper presents an approach that leverages recent work in manifold-learning on sequences to discover word clusters from language data, including both syntactic classes and semantic topics. From unlabeled data we form a smooth. low-dimensional feature space, where each word token is projected based on its underlying role as a function or content word. We then use this projection as additional input features to a linear-chain conditional random field trained on limited labeled training data. On standard part-of-speech tagging and Chinese word segmentation data sets we show as much as 14% error reduction due to the unlabeled data, and also statistically-significant improvements over a related semi-supervised sequence tagging method due to Miller et al.