Tied-mixture language modeling in continuous space

  • Authors:
  • Ruhi Sarikaya;Mohamed Afify;Brian Kingsbury

  • Affiliations:
  • IBM T.J. Watson Research Center, Yorktown Heights, NY;Orange Labs, Cairo, Egypt;IBM T.J. Watson Research Center, Yorktown Heights, NY

  • Venue:
  • NAACL '09 Proceedings of Human Language Technologies: The 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics
  • Year:
  • 2009
  • Deep neural network language models

    WLM '12 Proceedings of the NAACL-HLT 2012 Workshop: Will We Ever Really Replace the N-gram Model? On the Future of Language Modeling for HLT

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper presents a new perspective to the language modeling problem by moving the word representations and modeling into the continuous space. In a previous work we introduced Gaussian-Mixture Language Model (GMLM) and presented some initial experiments. Here, we propose Tied-Mixture Language Model (TMLM), which does not have the model parameter estimation problems that GMLM has. TMLM provides a great deal of parameter tying across words, hence achieves robust parameter estimation. As such, TMLM can estimate the probability of any word that has as few as two occurrences in the training data. The speech recognition experiments with the TMLM show improvement over the word trigram model.