Speech-Driven Facial Animation Using a Shared Gaussian Process Latent Variable Model

  • Authors:
  • Salil Deena;Aphrodite Galata

  • Affiliations:
  • School of Computer Science, University of Manchester, Manchester, United Kingdom M13 9PL;School of Computer Science, University of Manchester, Manchester, United Kingdom M13 9PL

  • Venue:
  • ISVC '09 Proceedings of the 5th International Symposium on Advances in Visual Computing: Part I
  • Year:
  • 2009

Quantified Score

Hi-index 0.01

Visualization

Abstract

In this work, synthesis of facial animation is done by modelling the mapping between facial motion and speech using the shared Gaussian process latent variable model. Both data are processed separately and subsequently coupled together to yield a shared latent space. This method allows coarticulation to be modelled by having a dynamical model on the latent space. Synthesis of novel animation is done by first obtaining intermediate latent points from the audio data and then using a Gaussian Process mapping to predict the corresponding visual data. Statistical evaluation of generated visual features against ground truth data compares favourably with known methods of speech animation. The generated videos are found to show proper synchronisation with audio and exhibit correct facial dynamics.