Visual object-action recognition: Inferring object affordances from human demonstration

  • Authors:
  • Hedvig Kjellström;Javier Romero;Danica Kragić

  • Affiliations:
  • Computational Vision and Active Perception Lab, Centre for Autonomous Systems School of Computer Science and Communication, KTH, SE-100 44 Stockholm, Sweden;Computational Vision and Active Perception Lab, Centre for Autonomous Systems School of Computer Science and Communication, KTH, SE-100 44 Stockholm, Sweden;Computational Vision and Active Perception Lab, Centre for Autonomous Systems School of Computer Science and Communication, KTH, SE-100 44 Stockholm, Sweden

  • Venue:
  • Computer Vision and Image Understanding
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper investigates object categorization according to function, i.e., learning the affordances of objects from human demonstration. Object affordances (functionality) are inferred from observations of humans using the objects in different types of actions. The intended application is learning from demonstration, in which a robot learns to employ objects in household tasks, from observing a human performing the same tasks with the objects. We present a method for categorizing manipulated objects and human manipulation actions in context of each other. The method is able to simultaneously segment and classify human hand actions, and detect and classify the objects involved in the action. This can serve as an initial step in a learning from demonstration method. Experiments show that the contextual information improves the classification of both objects and actions.