Lipreading with local spatiotemporal descriptors

  • Authors:
  • Guoying Zhao;Mark Barnard;Matti Pietikäinen

  • Affiliations:
  • Machine Vision Group, Infotech Oulu and Department of Electrical and Information Engineering, University of Oulu, Oulu, Finland;Machine Vision Group, University of Oulu, Oulu, Finland and Faculty of Engineering and Physical Sciences, University of Surrey, Guildford, UK;Machine Vision Group, Infotech Oulu and Department of Electrical and Information Engineering, University of Oulu, Oulu, Finland

  • Venue:
  • IEEE Transactions on Multimedia
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Visual speech information plays an important role in lipreading under noisy conditions or for listeners with a hearing impairment. In this paper, we present local spatiotemporal descriptors to represent and recognize spoken isolated phrases based solely on visual input. Spatiotemporal local binary patterns extracted from mouth regions are used for describing isolated phrase sequences. In our experiments with 817 sequences from ten phrases and 20 speakers, promising accuracies of 62% and 70% were obtained in speaker-independent and speaker-dependent recognition, respectively. In comparison with other methods on AVLetters database, the accuracy, 62.8%, of our method clearly outperforms the others. Analysis of the confusion matrix for 26 English letters shows the good clustering characteristics of visemes for the proposed descriptors. The advantages of our approach include local processing and robustness to monotonic gray-scale changes. Moreover, no error prone segmentation of moving lips is needed.