Adaptation of large vocabulary recognition system parameters

  • Authors:
  • L. Bahl;P. V. de Souza;D. Nahamoo;M. A. Picheny;S. Roukos

  • Affiliations:
  • Computer Sciences Department, IBM Thomas J. Watson Research Center, Yorktown Heights, NY;Computer Sciences Department, IBM Thomas J. Watson Research Center, Yorktown Heights, NY;Computer Sciences Department, IBM Thomas J. Watson Research Center, Yorktown Heights, NY;Computer Sciences Department, IBM Thomas J. Watson Research Center, Yorktown Heights, NY;Computer Sciences Department, IBM Thomas J. Watson Research Center, Yorktown Heights, NY

  • Venue:
  • ICASSP'92 Proceedings of the 1992 IEEE international conference on Acoustics, speech and signal processing - Volume 1
  • Year:
  • 1992

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper reports on a series of experiments in which the Hidden Markov Model baseforms and the language model probabilities were updated from spontaneously dictated speech captured during recognition sessions with the IBM Tangora system. The basic technique for baseform modification consisted of constructing new fenonic baseforms for all recognized words. To modify the language model probabilities, a simplified version of a cache language model was implemented. The word error rate across six talkers was 3.7%. Baseform adaptation reduced the average error rate to 3.5%, and employing the cache language model reduced the error rate to 3.2%. Combining both techniques further reduced the error rate to 3.1% - a respectable improvement over the original error rate, especially given that the system was speaker-trained prior to adaptation.