Speech-driven cartoon animation with emotions

  • Authors:
  • Yan Li;Feng Yu;Ying-Qing Xu;Eric Chang;Heung-Yeung Shum

  • Affiliations:
  • Microsoft Research China, Beijing, China;Tsinghua University, China;Microsoft Research China, Beijing, China;Microsoft Research China, Beijing, China;Microsoft Research China, Beijing, China

  • Venue:
  • MULTIMEDIA '01 Proceedings of the ninth ACM international conference on Multimedia
  • Year:
  • 2001

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we present a cartoon face animation system for multimedia HCI applications. We animate face cartoons not only from input speech, but also based on emotions derived from speech signal. Using a corpus of over 700 utterances from different speakers, we have trained SVMs (support vector machines) to recognize four categories of emotions: neutral, happiness, anger and sadness. Given each input speech phrase, we identify its emotion content as a mixture of all four emotions, rather than classifying it into a single emotion. Then, facial expressions are= generated from the recovered emotion for each phrase, by morphing different cartoon templates that correspond to various emotions. To ensure smooth transitions in the animation, we apply low-pass filtering to the recovered (and possibly jumpy) emotion sequence. Moreover, lip-syncing is applied to produce the lip movement from speech, by recovering a statistical audio-visual mapping. Experimental results demonstrate that cartoon animation sequences generated by our system are of good and convincing quality.