Statistical language models within the algebra of weighted rational languages

  • Authors:
  • Thomas Hanneforth;Kay-Michael Würzner

  • Affiliations:
  • University of Potsdam;University of Potsdam

  • Venue:
  • Acta Cybernetica
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Statistical language models are an important tool in natural language processing. They represent prior knowledge about a certain language which is usually gained from a set of samples called a corpus. In this paper, we present a novel way of creating N-gram language models using weighted finite automata. The construction of these models is formalised within the algebra underlying weighted finite automata and expressed in terms of weighted rational languages and transductions. Besides the algebra we make use of five special constant weighted transductions which rely only on the alphabet and the model parameter N. In addition, we discuss efficient implementations of these transductions in terms of virtual constructions.