Articulated Multi-body Tracking under Egomotion

  • Authors:
  • Stephan Gammeter;Andreas Ess;Tobias Jäggli;Konrad Schindler;Bastian Leibe;Luc Gool

  • Affiliations:
  • ETH Zürich, ;ETH Zürich, ;ETH Zürich, ;ETH Zürich, ;ETH Zürich, and RWTH Aachen, ;ETH Zürich, and KU Leuven, IBBT,

  • Venue:
  • ECCV '08 Proceedings of the 10th European Conference on Computer Vision: Part II
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we address the problem of 3D articulated multi-person tracking in busy street scenes from a moving, human-level observer. In order to handle the complexity of multi-person interactions, we propose to pursue a two-stage strategy. A multi-body detection-based tracker first analyzes the scene and recovers individual pedestrian trajectories, bridging sensor gaps and resolving temporary occlusions. A specialized articulated tracker is then applied to each recovered pedestrian trajectory in parallel to estimate the tracked person's precise body pose over time. This articulated tracker is implemented in a Gaussian Process framework and operates on global pedestrian silhouettes using a learned statistical representation of human body dynamics. We interface the two tracking levels through a guided segmentation stage, which combines traditional bottom-up cues with top-down information from a human detector and the articulated tracker's shape prediction. We show the proposed approach's viability and demonstrate its performance for articulated multi-person tracking on several challenging video sequences of a busy inner-city scenario.