Face active appearance modeling and speech acoustic information to recover articulation

  • Authors:
  • Athanassios Katsamanis;George Papandreou;Petros Maragos

  • Affiliations:
  • School of Electrical and Computer Engineering, National Technical University of Athens, Athens, Greece;School of Electrical and Computer Engineering, National Technical University of Athens, Athens, Greece;School of Electrical and Computer Engineering, National Technical University of Athens, Athens, Greece

  • Venue:
  • IEEE Transactions on Audio, Speech, and Language Processing - Special issue on multimodal processing in speech-based interactions
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

We are interested in recovering aspects of vocal tract's geometry and dynamics from speech, a problem referred to as speech inversion. Traditional audio-only speech inversion techniques are inherently ill-posed since the same speech acoustics can be produced by multiple articulatory configurations. To alleviate the ill-posedness of the audio-only inversion process, we propose an inversion scheme which also exploits visual information from the speaker's face, The complex audiovisual-to-articulatory mapping is approximated by an adaptive piecewise linear model. Model switching is governed by a Markovian discrete process which captures articulatory dynamic information. Each constituent linear mapping is effectively estimated via canonical correlation analysis. In the described multimodal context, we investigate alternative fusion schemes which allow interaction between the audio and visual modalities at various synchronization levels. For facial analysis, we employ active appearance models (AAMs) and demonstrate fully automatic face tracking and visual feature extraction. Using the AAM features in conjunction with audio features such as Mel frequency cepstral coefficients (MFCCs) or line spectral frequencies (LSFs) leads to effective estimation of the trajectories followed by certain points of interest in the speech production system. We report experiments on the QSMT and MOCHA databases which contain audio, video, and electromagnetic articulography data recorded in parallel. The results show that exploiting both audio and visual modalities in a multistream hidden Markov model based scheme clearly improves performance relative to either audio or visual-only estimation.