Fast and Robust Face Tracking for Analyzing Multiparty Face-to-Face Meetings

  • Authors:
  • Kazuhiro Otsuka;Junji Yamato

  • Affiliations:
  • NTT Communication Science Labs., , Atsugi-shi, Japan 243-0198;NTT Communication Science Labs., , Atsugi-shi, Japan 243-0198

  • Venue:
  • MLMI '08 Proceedings of the 5th international workshop on Machine Learning for Multimodal Interaction
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper presents a novel face tracker and verifies its effectiveness for analyzing group meetings. In meeting scene analysis, face direction is an important clue for assessing the visual attention of meeting participants. The face tracker, called STCTracker (Sparse Template Condensation Tracker), estimates face position and pose by matching face templates in the framework of a particle filter. STCTracker is robust against large head rotation, up to ±60 degrees in the horizontal direction, with relatively small mean deviation error. Also, it can track multiple faces simultaneously in real-time by utilizing a modern GPU (Graphics Processing Unit), e.g. 6 faces at about 28 frames/second on a single PC. Also, it can automatically build 3-D face templates upon initialization of the tracker. This paper evaluates the tracking errors and verifies the effectiveness of STCTracker for meeting scene analysis, in terms of conversation structures, gaze directions, and the structure of cross-modal interactions involving head gestures and utterances. Experiments confirm that STCTracker can basically match the performance of from the user-unfriendly magnetic-sensor-based motion capture system.