Multiple Parallel Vision-Based Recognition in a Real-Time Framework for Human-Robot-Interaction Scenarios

  • Authors:
  • Tobias Rehrl;Alexander Bannat;Jürgen Gast;Frank Wallhoff;Gerhard Rigoll;Christoph Mayer;Zadid Riaz;Bernd Radig;Stefan Sosnowski;Kolja Kühnlenz

  • Affiliations:
  • -;-;-;-;-;-;-;-;-;-

  • Venue:
  • ACHI '10 Proceedings of the 2010 Third International Conference on Advances in Computer-Human Interactions
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

Every day human communication relies on a large number of different communication mechanisms like spoken language, facial expressions, body pose and gestures, allowing humans to pass large amounts of information in short time. In contrast, traditional human-machine communication is often unintuitive and requires specifically trained personal. In this paper, we present a real-time capable framework that recognizes traditional visual human communication signals in order to establish a more intuitive human-machine interaction. Humans rely on the interaction partner’s face for identification, which helps them to adapt to the interaction partner and utilize context information. Head gestures (head nodding and head shaking) are a convenient way to show agreement or disagreement. Facial expressions give evidence about the interaction partners’ emotional state and hand gestures are a fast way of passing simple commands. The recognition of all interaction queues is performed in parallel, enabled by a shared memory implementation.