An improved automatic lipreading system to enhance speech recognition
CHI '88 Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
Neural Networks for Pattern Recognition
Neural Networks for Pattern Recognition
ECCV '98 Proceedings of the 5th European Conference on Computer Vision-Volume II - Volume II
On merging hidden Markov models with deformable templates
ICIP '95 Proceedings of the 1995 International Conference on Image Processing (Vol. 3)-Volume 3 - Volume 3
Lipreading method using color extraction method and eigenspace technique
Systems and Computers in Japan
Non-Euclidean c-means clustering algorithms
Intelligent Data Analysis
Accurate and quasi-automatic lip tracking
IEEE Transactions on Circuits and Systems for Video Technology
Integrated Computer-Aided Engineering
Integrated Computer-Aided Engineering
Talking Agents: A distributed architecture for interactive artistic installations
Integrated Computer-Aided Engineering
Integration of emerging computer technologies for an efficient image sequences analysis
Integrated Computer-Aided Engineering
Human automatic detection and tracking for outdoor video
Integrated Computer-Aided Engineering
Hi-index | 0.00 |
An automatic lip-reading system is among assistive technologies for hearing impaired or elderly people. We can imagine, for example, a dependent person ordering a machine with an easy lip movement or by a simple visemes (visual phoneme) pronunciation. A lip-reading system is decomposed into three subsystems: a lip localization subsystem, then a feature extracting subsystem, followed by a classification system that maps feature vectors to visemes. The major difficulty in a lip-reading system is the extraction of the visual speech descriptors. In fact, to ensure this task it is necessary to carry out an automatic localization and tracking of the labial gestures. We present, in this paper, a new automatic approach for lip POI localization and feature extraction on a speaker's face based on mouth color information and a geometrical model of the lips. The extracted visual information is then classified in order to recognize the uttered viseme. We have developed our Automatic Lip Feature Extraction prototype (ALiFE). ALiFE prototype is evaluated for multiple speakers under natural conditions. Experiments include a group of French visemes for different speakers. Results revealed that our system recognizes 94.64% of the tested French visemes.