Dynamic units of visual speech

  • Authors:
  • Sarah L. Taylor;Moshe Mahler;Barry-John Theobald;Iain Matthews

  • Affiliations:
  • University of East Anglia, Norwich, England;Disney Research, Pittsburgh;University of East Anglia, Norwich, England;Disney Research, Pittsburgh

  • Venue:
  • EUROSCA'12 Proceedings of the 11th ACM SIGGRAPH / Eurographics conference on Computer Animation
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

We present a new method for generating a dynamic, concatenative, unit of visual speech that can generate realistic visual speech animation. We redefine visemes as temporal units that describe distinctive speech movements of the visual speech articulators. Traditionally visemes have been surmized as the set of static mouth shapes representing clusters of contrastive phonemes (e.g. /p, b, m/, and /f, v/). In this work, the motion of the visual speech articulators are used to generate discrete, dynamic visual speech gestures. These gestures are clustered, providing a finite set of movements that describe visual speech, the visemes. Dynamic visemes are applied to speech animation by simply concatenating viseme units. We compare to static visemes using subjective evaluation. We find that dynamic visemes are able to produce more accurate and visually pleasing speech animation given phonetically annotated audio, reducing the amount of time that an animator needs to spend manually refining the animation.