Lip-synching using speaker-specific articulation, shape and appearance models

  • Authors:
  • Gérard Bailly;Oxana Govokhina;Frédéric Elisei;Gaspard Breton

  • Affiliations:
  • Department of Speech and Cognition, GIPSA-Lab, CNRS & Grenoble University, Saint Martin d'Hères cedex, France;Department of Speech and Cognition, GIPSA-Lab, CNRS & Grenoble University, Saint Martin d'Hères cedex, France and Orange Labs, Cesson-Sévigné, France;Department of Speech and Cognition, GIPSA-Lab, CNRS & Grenoble University, Saint Martin d'Hères cedex, France;Orange Labs, Cesson-Sévigné, France

  • Venue:
  • EURASIP Journal on Audio, Speech, and Music Processing - Special issue on animating virtual speakers or singers from audio: Lip-synching facial animation
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

We describe here the control, shape and appearance models that are built using an original photogrammetric method to capture characteristics of speaker-specific facial articulation, anatomy, and texture. Two original contributions are put forward here: the trainable trajectory formation model that predicts articulatory trajectories of a talking face from phonetic input and the texture model that computes a texture for each 3D facial shape according to articulation. Using motion capture data from different speakers and module-specific evaluation procedures, we show here that this cloning system restores detailed idiosyncrasies and the global coherence of visible articulation. Results of a subjective evaluation of the global system with competing trajectory formation models are further presented and commented.