Multi-level Particle Filter Fusion of Features and Cues for Audio-Visual Person Tracking

  • Authors:
  • Keni Bernardin;Tobias Gehrig;Rainer Stiefelhagen

  • Affiliations:
  • Interactive Systems Lab Institut für Theoretische Informatik, Universität Karlsruhe, Karlsruhe, Germany 76131;Interactive Systems Lab Institut für Theoretische Informatik, Universität Karlsruhe, Karlsruhe, Germany 76131;Interactive Systems Lab Institut für Theoretische Informatik, Universität Karlsruhe, Karlsruhe, Germany 76131

  • Venue:
  • Multimodal Technologies for Perception of Humans
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, two multimodal systems for the tracking of multiple users in smart environments are presented. The first is a multi-view particle filter tracker using foreground, color and special upper body detection and person region features. The other is a wide angle overhead view person tracker relying on foreground segmentation and model-based blob tracking. Both systems are completed by a joint probabilistic data association filter-based source localizer using the input from several microphone arrays. While the first system fuses audio and visual cues at the feature level, the second one incorporates them at the decision level using state-based heuristics.The systems are designed to estimate the 3D scene locations of room occupants and are evaluated based on their precision in estimating person locations, their accuracy in recognizing person configurations and their ability to consistently keep track identities over time.The trackers are extensively tested and compared, for each separate modality and for the combined modalities, on the CLEAR 2007 Evaluation Database.