Recognizing Gestures for Virtual and Real World Interaction

  • Authors:
  • David Demirdjian;Chenna Varri

  • Affiliations:
  • MIT CSAIL, Cambridge, USA 02142 and Toyota Research Institute, Cambridge, USA 02142;Toyota Research Institute, Cambridge, USA 02142

  • Venue:
  • ICVS '09 Proceedings of the 7th International Conference on Computer Vision Systems: Computer Vision Systems
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we present a vision-based system that estimates the pose of users as well as the gestures they perform in real time. This system allow users to interact naturally with an application (virtual reality, gaming) or a robot. The main components of our system are a 3D upper-body tracker, which estimates human body pose in real-time from a stereo sensor and a gesture recognizer, which classifies output from temporal tracker into gesture classes. The main novelty of our system is the bag-of-features representation for temporal sequences. This representation, though simple, proves to be surprisingly powerful and able to implicitly learn sequence dynamics. Based on this representation, a multi-class classifier, treating the bag of features as the feature vector is applied to estimate the corresponding gesture class. We show with experiments performed on a HCI gesture dataset that our method performs better than state-of-the-art algorithms and has some nice generalization properties. Finally, we describe virtual and real world applications, in which our system was integrated for multimodal interaction.