A biological and real-time framework for hand gestures and head poses

  • Authors:
  • Mario Saleiro;Miguel Farrajota;Kasim Terzić;João M. F. Rodrigues;J. M. Hans du Buf

  • Affiliations:
  • Vision Laboratory, LARSyS, University of the Algarve, Faro, Portugal;Vision Laboratory, LARSyS, University of the Algarve, Faro, Portugal;Vision Laboratory, LARSyS, University of the Algarve, Faro, Portugal;Vision Laboratory, LARSyS, University of the Algarve, Faro, Portugal;Vision Laboratory, LARSyS, University of the Algarve, Faro, Portugal

  • Venue:
  • UAHCI'13 Proceedings of the 7th international conference on Universal Access in Human-Computer Interaction: design methods, tools, and interaction techniques for eInclusion - Volume Part I
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Human-robot interaction is an interdisciplinary research area that aims at the development of social robots. Since social robots are expected to interact with humans and understand their behavior through gestures and body movements, cognitive psychology and robot technology must be integrated. In this paper we present a biological and real-time framework for detecting and tracking hands and heads. This framework is based on keypoints extracted by means of cortical V1 end-stopped cells. Detected keypoints and the cells' responses are used to classify the junction type. Through the combination of annotated keypoints in a hierarchical, multi-scale tree structure, moving and deformable hands can be segregated and tracked over time. By using hand templates with lines and edges at only a few scales, a hand's gestures can be recognized. Head tracking and pose detection are also implemented, which can be integrated with detection of facial expressions in the future. Through the combinations of head poses and hand gestures a large number of commands can be given to a robot.