Real-Time American Sign Language Recognition Using Desk and Wearable Computer Based Video
IEEE Transactions on Pattern Analysis and Machine Intelligence
A System for Person-Independent Hand Posture Recognition against Complex Backgrounds
IEEE Transactions on Pattern Analysis and Machine Intelligence
A Multi-Class Pattern Recognition System for Practical Finger Spelling Translation
ICMI '02 Proceedings of the 4th IEEE International Conference on Multimodal Interfaces
Virtual 3-D interface system via hand motion recognition from two cameras
IEEE Transactions on Systems, Man, and Cybernetics, Part A: Systems and Humans
Sign Language Recognition: Working with Limited Corpora
UAHCI '09 Proceedings of the 5th International Conference on Universal Access in Human-Computer Interaction. Part III: Applications and Services
Large lexicon detection of sign language
HCI'07 Proceedings of the 2007 IEEE international conference on Human-computer interaction
Pronunciation clustering and modeling of variability for appearance-based sign language recognition
GW'05 Proceedings of the 6th international conference on Gesture in Human-Computer Interaction and Simulation
PR'05 Proceedings of the 27th DAGM conference on Pattern Recognition
Non-manual cues in automatic sign language recognition
Personal and Ubiquitous Computing
International Journal of Computational Vision and Robotics
Hi-index | 0.00 |
In this paper, we present how appearance-based features can be used for the recognition of words in American sign language (ASL) from a video stream. The features are extracted without any segmentation or tracking of the hands or head of the signer, which avoids possible errors in the segmentation step. Experiments are performed on a database that consists of 10 words in ASL with 110 utterances in total. These data are extracted from a publicly available collection of videos and can therefore be used by other research groups. The video streams of two stationary cameras are used for classification, but we observe that one camera alone already leads to sufficient accuracy. Hidden Markov Models and the leaving one out method are employed for training and classification. Using the simple appearance-based features, we achieve an error rate of 7%. About half of the remaining errors are due to words that are visually different from all other utterances.