A new class of fenonic Markov word models for large vocabulary continuous speech recognition

  • Authors:
  • L. R. Bahl;J. R. Bellegarda;P. V. deSouza;P. S. Gopalakrishnan;D. Nahamoo;M. A. Picheny

  • Affiliations:
  • IBM, Thomas Watson Res. Center, Yorktown Heights, NY, USA;IBM, Thomas Watson Res. Center, Yorktown Heights, NY, USA;IBM, Thomas Watson Res. Center, Yorktown Heights, NY, USA;IBM, Thomas Watson Res. Center, Yorktown Heights, NY, USA;IBM, Thomas Watson Res. Center, Yorktown Heights, NY, USA;IBM, Thomas Watson Res. Center, Yorktown Heights, NY, USA

  • Venue:
  • ICASSP '91 Proceedings of the Acoustics, Speech, and Signal Processing, 1991. ICASSP-91., 1991 International Conference
  • Year:
  • 1991

Quantified Score

Hi-index 0.00

Visualization

Abstract

A technique for constructing hidden Markov models for the acoustic representation of words is described. The models, built from combinations of acoustically based subword units called fenones, are derived automatically from one or more sample utterances of words. They are more flexible than previously reported fenone-based word models and lead to an improved capability of modeling variations in pronunciation. In addition, their construction is simplified, because it can be done using the well-known forward-backward algorithm for the parameter estimation of hidden Markov models. Experimental results obtained on a 5000-word vocabulary continuous speech recognition task are presented to illustrate some of the benefits associated with the new models. Multonic baseforms resulted in a reduction of 16% in the average error rate obtained for ten speakers.