Boosted Audio-Visual HMM for Speech Reading

  • Authors:
  • Pei Yin;Irfan Essa;James M. Rehg

  • Affiliations:
  • -;-;-

  • Venue:
  • AMFG '03 Proceedings of the IEEE International Workshop on Analysis and Modeling of Faces and Gestures
  • Year:
  • 2003

Quantified Score

Hi-index 0.02

Visualization

Abstract

We propose a new approach for combining acoustic andvisual measurements to aid in recognizing lip shapes of aperson speaking. Our method relies on computing the maximumlikelihoods of (a) HMM used to model phonemes fromthe acoustic signal, and (b) HMM used to model visual featuresmotions from video. One significant addition in thiswork is the dynamic analysis with features selected by Ad-aBoost,on the basis of their discriminant ability. This formof integration, leading to boosted HMM, permits AdaBoostto find the best features first, and then uses HMM to exploitdynamic information inherent in the signal.