Scene semantics from long-term observation of people

  • Authors:
  • Vincent Delaitre;David F. Fouhey;Ivan Laptev;Josef Sivic;Abhinav Gupta;Alexei A. Efros

  • Affiliations:
  • INRIA/École Normale Supérieure, Paris, France;Carnegie Mellon University;INRIA/École Normale Supérieure, Paris, France;INRIA/École Normale Supérieure, Paris, France;Carnegie Mellon University;INRIA/École Normale Supérieure, Paris, France, Carnegie Mellon University

  • Venue:
  • ECCV'12 Proceedings of the 12th European conference on Computer Vision - Volume Part VI
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

Our everyday objects support various tasks and can be used by people for different purposes. While object classification is a widely studied topic in computer vision, recognition of object function, i.e., what people can do with an object and how they do it, is rarely addressed. In this paper we construct a functional object description with the aim to recognize objects by the way people interact with them. We describe scene objects (sofas, tables, chairs) by associated human poses and object appearance. Our model is learned discriminatively from automatically estimated body poses in many realistic scenes. In particular, we make use of time-lapse videos from YouTube providing a rich source of common human-object interactions and minimizing the effort of manual object annotation. We show how the models learned from human observations significantly improve object recognition and enable prediction of characteristic human poses in new scenes. Results are shown on a dataset of more than 400,000 frames obtained from 146 time-lapse videos of challenging and realistic indoor scenes.