Combining Topic Information and Structure Information in a Dynamic Language Model

  • Authors:
  • Pascal Wiggers;Leon Rothkrantz

  • Affiliations:
  • Man---Machine Interaction Group, Delft University of Technology, Delft, The Netherlands 2628;Man---Machine Interaction Group, Delft University of Technology, Delft, The Netherlands 2628

  • Venue:
  • TSD '09 Proceedings of the 12th International Conference on Text, Speech and Dialogue
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

We present a language model implemented with dynamic Bayesian networks that combines topic information and structure information to capture long distance dependencies between the words in a text while maintaining the robustness of standard n -gram models. We show that the model is an extension of sentence level mixture models, thereby providing a Bayesian explanation for these models. We describe a procedure for unsupervised training of the model. Experiments show that it reduces perplexity by 13% compared to an interpolated trigram.