Nonlinear dynamic shape and appearance models for facial motion tracking

  • Authors:
  • Chan-Su Lee;Ahmed Elgammal;Dimitris Metaxas

  • Affiliations:
  • Rutgers University, Piscataway, NJ;Rutgers University, Piscataway, NJ;Rutgers University, Piscataway, NJ

  • Venue:
  • PSIVT'07 Proceedings of the 2nd Pacific Rim conference on Advances in image and video technology
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

We present a framework for tracking large facial deformations using nonlinear dynamic shape and appearance model based upon local motion estimation. Local facial deformation estimation based on a given single template fails to track large facial deformations due to significant appearance variations. A nonlinear generative model that uses low dimensional manifold representation provides adaptive facial appearance templates depending upon the movement of the facial motion state and the expression type. The proposed model provides a generative model for Bayesian tracking of facial motions using particle filtering with simultaneous estimation of the expression type. We estimate the geometric transformation and the global deformation using the generative model. The appearance templates from the global model then estimate local deformation based on thin-plate spline parameters.