Robust AAM-based audio-visual speech recognition against face direction changes

  • Authors:
  • Yuto Komai;Nan Yang;Tetsuya Takiguchi;Yasuo Ariki

  • Affiliations:
  • Kobe University, Kobe, Japan;Kobe University, Kobe, Japan;Kobe University, Kobe, Japan;Kobe University, Kobe, Japan

  • Venue:
  • Proceedings of the 20th ACM international conference on Multimedia
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

As one of the techniques for robust speech recognition under noisy environments, audio-visual speech recognition (AVSR) using lip dynamic scene information together with audio information is attracting attention, and the research has advanced in recent years. However, in visual speech recognition (VSR), when a face turns sideways, the shape of the lip as viewed from the camera changes and the recognition accuracy degrades significantly. Therefore, many of the conventional VSR methods are limited to situations in which the face is viewed from the front. This paper proposes a VSR method to convert faces viewed from various directions into faces that are viewed from the front using Active Appearance Models (AAM). In the experiment, even when the face direction changes about 30 degrees relative to a frontal view, the recognition accuracy improved significantly.