Lipreading from color video

  • Authors:
  • G. I. Chiou;Jenq-Neng Hwang

  • Affiliations:
  • Dept. of Electr. Eng., Washington Univ., Seattle, WA;-

  • Venue:
  • IEEE Transactions on Image Processing
  • Year:
  • 1997

Quantified Score

Hi-index 0.01

Visualization

Abstract

We have designed and implemented a lipreading system that recognizes isolated words using only color video of human lips (without acoustic data). The system performs video recognition using “snakes” to extract visual features of geometric space, Karhunen-Loeve transform (KLT) to extract principal components in the color eigenspace, and hidden Markov models (HMM's) to recognize the combined visual features sequences. With the visual information alone, we were able to achieve 94% accuracy for ten isolated words