Learning context for collective activity recognition

  • Authors:
  • Wongun Choi;K. Shahid;S. Savarese

  • Affiliations:
  • Dept. of Electr. & Comput. Eng., Univ. of Michigan, Ann Arbor, MI, USA;Dept. of Electr. & Comput. Eng., Univ. of Michigan, Ann Arbor, MI, USA;Dept. of Electr. & Comput. Eng., Univ. of Michigan, Ann Arbor, MI, USA

  • Venue:
  • CVPR '11 Proceedings of the 2011 IEEE Conference on Computer Vision and Pattern Recognition
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper we present a framework for the recognition of collective human activities. A collective activity is defined or reinforced by the existence of coherent behavior of individuals in time and space. We call such coherent behavior 'Crowd Context'. Examples of collective activities are "queuing in a line" or "talking". Following, we propose to recognize collective activities using the crowd context and introduce a new scheme for learning it automatically. Our scheme is constructed upon a Random Forest structure which randomly samples variable volume spatio-temporal regions to pick the most discriminating attributes for classification. Unlike previous approaches, our algorithm automatically finds the optimal configuration of spatio-temporal bins, over which to sample the evidence, by randomization. This enables a methodology for modeling crowd context. We employ a 3D Markov Random Field to regularize the classification and localize collective activities in the scene. We demonstrate the flexibility and scalability of the proposed framework in a number of experiments and show that our method outperforms state-of-the art action classification techniques.