Visual Sign Language Recognition

  • Authors:
  • Eun-Jung Holden;Robyn A. Owens

  • Affiliations:
  • -;-

  • Venue:
  • Proceedings of the 10th International Workshop on Theoretical Foundations of Computer Vision: Multi-Image Analysis
  • Year:
  • 2000

Quantified Score

Hi-index 0.00

Visualization

Abstract

Automatic gesture recognition systems generally require two separate processes: a motion sensing process where some motion features are extracted from the visual input; and a classification process where the features are recognised as gestures. We have developed the Hand Motion Understanding (HMU) system that uses the combination of a 3D model-based hand tracker for motion sensing and an adaptive fuzzy expert system for motion classification. The HMU system understands static and dynamic hand signs of the Australian Sign Language (Auslan). This paper presents the hand tracker that extracts 3D hand configuration data with 21 degrees-of-freedom (DOFs) from a 2D image sequence that is captured from a single viewpoint, with the aid of a colour-coded glove. Then the temporal sequence of 3D hand configurations detected by the tracker is recognised as a sign by an adaptive fuzzy expert system. The HMU system was evaluated with 22 static and dynamic signs. Before training the HMU system achieved 91% recognition, and after training it achieved over 95% recognition.