A framework for recognizing the simultaneous aspects of American sign language
Computer Vision and Image Understanding - Modeling people toward vision-based underatanding of a person's shape, appearance, and movement
A systematic comparison of various statistical alignment models
Computational Linguistics
Automatic Sign Language Analysis: A Survey and the Future beyond Lexical Meaning
IEEE Transactions on Pattern Analysis and Machine Intelligence
Statistical phrase-based translation
NAACL '03 Proceedings of the 2003 Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology - Volume 1
Acoustic Modeling of Speaking Styles and Emotional Expressions in HMM-Based Speech Synthesis
IEICE - Transactions on Information and Systems
Hidden Conditional Random Fields for Gesture Recognition
CVPR '06 Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Volume 2
Rapid Signer Adaptation for Isolated Sign Language Recognition
CVPRW '06 Proceedings of the 2006 Conference on Computer Vision and Pattern Recognition Workshop
Real Time Large Vocabulary Continuous Sign Language Recognition Based on OP/Viterbi Algorithm
ICPR '06 Proceedings of the 18th International Conference on Pattern Recognition - Volume 03
Details of the Nitech HMM-Based Speech Synthesis System for the Blizzard Challenge 2005
IEICE - Transactions on Information and Systems
Taiwan sign language (TSL) recognition based on 3D data and neural networks
Expert Systems with Applications: An International Journal
Enhancing a Sign Language Translation System with Vision-Based Features
Gesture-Based Human-Computer Interaction and Simulation
Robust speaker-adaptive HMM-based text-to-speech synthesis
IEEE Transactions on Audio, Speech, and Language Processing
Spoken Spanish generation from sign language
Interacting with Computers
Persian sign language (PSL) recognition using wavelet transform and neural networks
Expert Systems with Applications: An International Journal
Design, development and field evaluation of a Spanish into sign language translation system
Pattern Analysis & Applications
Methodology for developing an advanced communications system for the Deaf in a new domain
Knowledge-Based Systems
Hi-index | 12.05 |
This paper describes the development of LSESpeak, a spoken Spanish generator for Deaf people. This system integrates two main tools: a sign language into speech translation system and an SMS (Short Message Service) into speech translation system. The first tool is made up of three modules: an advanced visual interface (where a deaf person can specify a sequence of signs), a language translator (for generating the sequence of words in Spanish), and finally, an emotional text to speech (TTS) converter to generate spoken Spanish. The visual interface allows a sign sequence to be defined using several utilities. The emotional TTS converter is based on Hidden Semi-Markov Models (HSMMs) permitting voice gender, type of emotion, and emotional strength to be controlled. The second tool is made up of an SMS message editor, a language translator and the same emotional text to speech converter. Both translation tools use a phrase-based translation strategy where translation and target language models are trained from parallel corpora. In the experiments carried out to evaluate the translation performance, the sign language-speech translation system reported a 96.45 BLEU and the SMS-speech system a 44.36 BLEU in a specific domain: the renewal of the Identity Document and Driving License. In the evaluation of the emotional TTS, it is important to highlight the improvement in the naturalness thanks to the morpho-syntactic features, and the high flexibility provided by HSMMs when generating different emotional strengths.