Style learning and transferring for facial animation editing

  • Authors:
  • Xiaohan Ma;Binh Huy Le;Zhigang Deng

  • Affiliations:
  • University of Houston, Houston, TX;University of Houston, Houston, TX;University of Houston, Houston, TX

  • Venue:
  • Proceedings of the 2009 ACM SIGGRAPH/Eurographics Symposium on Computer Animation
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Most of current facial animation editing techniques are frame-based approaches (i.e., manually edit one keyframe every several frames), which is ineffective, time-consuming, and prone to editing inconsistency. In this paper, we present a novel facial editing style learning framework that is able to learn a constraint-based Gaussian Process model from a small number of facial-editing pairs, and then it can be effectively applied to automate the editing of the remaining facial animation frames or transfer editing styles between different animation sequences. Comparing with the state of the art, multiresolution-based mesh sequence editing technique, our approach is more flexible, powerful, and adaptive. Our approach can dramatically reduce the manual efforts required by most of current facial animation editing approaches.