Dynamics of facial expression extracted automatically from video

  • Authors:
  • Gwen Littlewort;Marian Stewart Bartlett;Ian Fasel;Joshua Susskind;Javier Movellan

  • Affiliations:
  • Institute for Neural Computation, University of California, Diego San Diego, CA 92093-0523, USA;Institute for Neural Computation, University of California, Diego San Diego, CA 92093-0523, USA;Institute for Neural Computation, University of California, Diego San Diego, CA 92093-0523, USA;Institute for Neural Computation, University of California, Diego San Diego, CA 92093-0523, USA;Institute for Neural Computation, University of California, Diego San Diego, CA 92093-0523, USA

  • Venue:
  • Image and Vision Computing
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

We present a systematic comparison of machine learning methods applied to the problem of fully automatic recognition of facial expressions, including AdaBoost, support vector machines, and linear discriminant analysis. Each video-frame is first scanned in real-time to detect approximately upright-frontal faces. The faces found are scaled into image patches of equal size, convolved with a bank of Gabor energy filters, and then passed to a recognition engine that codes facial expressions into 7 dimensions in real time: neutral, anger, disgust, fear, joy, sadness, surprise. We report results on a series of experiments comparing spatial frequency ranges, feature selection techniques, and recognition engines. Best results were obtained by selecting a subset of Gabor filters using AdaBoost and then training Support Vector Machines on the outputs of the filters selected by AdaBoost. The generalization performance to new subjects for a 7-way forced choice was 93% or more correct on two publicly available datasets, the best performance reported so far on these datasets. The outputs of the classifier change smoothly as a function of time and thus can be used for unobtrusive expression dynamics capture. We developed an end-to-end system that provides facial expression codes at 24 frames per second and animates a computer-generated character. In real-time this expression mirror operates down to resolutions of 16 pixels from eye to eye. We also applied the system to fully automated facial action coding.