Extracting Moving People from Internet Videos

  • Authors:
  • Juan Carlos Niebles;Bohyung Han;Andras Ferencz;Li Fei-Fei

  • Affiliations:
  • Princeton University, Princeton, USA and Universidad del Norte, Colombia;Mobileye Vision Technologies, Princeton, USA;Mobileye Vision Technologies, Princeton, USA;Princeton University, Princeton, USA

  • Venue:
  • ECCV '08 Proceedings of the 10th European Conference on Computer Vision: Part IV
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

We propose a fully automatic framework to detect and extract arbitrary human motion volumes from real-world videos collected from YouTube. Our system is composed of two stages. A person detector is first applied to provide crude information about the possible locations of humans. Then a constrained clustering algorithm groups the detections and rejects false positives based on the appearance similarity and spatio-temporal coherence. In the second stage, we apply a top-down pictorial structure model to complete the extraction of the humans in arbitrary motion. During this procedure, a density propagation technique based on a mixture of Gaussians is employed to propagate temporal information in a principled way. This method reduces greatly the search space for the measurement in the inference stage. We demonstrate the initial success of this framework both quantitatively and qualitatively by using a number of YouTube videos.