Learning to interpret pointing gestures with a time-of-flight camera

  • Authors:
  • David Droeschel;Jörg Stückler;Sven Behnke

  • Affiliations:
  • University of Bonn, Bonn, Germany;University of Bonn, Bonn, Germany;University of Bonn, Bonn, Germany

  • Venue:
  • Proceedings of the 6th international conference on Human-robot interaction
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Pointing gestures are a common and intuitive way to draw somebody's attention to a certain object. While humans can easily interpret robot gestures, the perception of human behavior using robot sensors is more difficult. In this work, we propose a method for perceiving pointing gestures using a Time-of-Flight (ToF) camera. To determine the intended pointing target, frequently the line between a person's eyes and hand is assumed to be the pointing direction. However, since people tend to keep the line-of-sight free while they are pointing, this simple approximation is inadequate. Moreover, depending on the distance and angle to the pointing target, the line between shoulder and hand or elbow and hand may yield better interpretations of the pointing direction. In order to achieve a better estimate, we extract a set of body features from depth and amplitude images of a ToF camera and train a model of pointing directions using Gaussian Process Regression. We evaluate the accuracy of the estimated pointing direction in a quantitative study. The results show that our learned model achieves far better accuracy than simple criteria like head-hand, shoulder-hand, or elbow-hand line.