Multi-Class Composite N-gram language model for spoken language processing using multiple word clusters

  • Authors:
  • Hirofumi Yamamoto;Shuntaro Isogai;Yoshinori Sagisaka

  • Affiliations:
  • ATR SLT, Soraku-gun, Kyoto-fu, Japan;Waseda University, Shinjuku-ku, Tokyo-to, Japan;GITI / ATR SLT, Shinjuku-ku, Tokyo-to, Japan

  • Venue:
  • ACL '01 Proceedings of the 39th Annual Meeting on Association for Computational Linguistics
  • Year:
  • 2001

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, a new language model, the Multi-Class Composite N-gram, is proposed to avoid a data sparseness problem for spoken language in that it is difficult to collect training data. The Multi-Class Composite N-gram maintains an accurate word prediction capability and reliability for sparse data with a compact model size based on multiple word clusters, called Multi-Classes. In the Multi-Class, the statistical connectivity at each position of the N-grams is regarded as word attributes, and one word cluster each is created to represent the positional attributes. Furthermore, by introducing higher order word N-grams through the grouping of frequent word successions, Multi-Class N-grams are extended to Multi-Class Composite N-grams. In experiments, the Multi-Class Composite N-grams result in 9.5% lower perplexity and a 16% lower word error rate in speech recognition with a 40% smaller parameter size than conventional word 3-grams.