Integrating simultaneous input from speech, gaze, and hand gestures
Intelligent multimedia interfaces
Natural language with integrated deictic and graphic gestures
Readings in intelligent user interfaces
A robust selection system using real-time multi-modal user-agent interactions
IUI '99 Proceedings of the 4th international conference on Intelligent user interfaces
Mutual disambiguation of recognition errors in a multimodel architecture
Proceedings of the SIGCHI conference on Human Factors in Computing Systems
“Put-that-there”: Voice and gesture at the graphics interface
SIGGRAPH '80 Proceedings of the 7th annual conference on Computer graphics and interactive techniques
Using eye movements to determine referents in a spoken dialogue system
Proceedings of the 2001 workshop on Perceptive user interfaces
A pen and speech-based storytelling system for Chinese children
Computers in Human Behavior
Artificial Intelligence Review
Gaze as a supplementary modality for interacting with ambient intelligence environments
UAHCI'07 Proceedings of the 4th international conference on Universal access in human-computer interaction: ambient interaction
A multimodal 3D storytelling system for Chinese children
Edutainment'07 Proceedings of the 2nd international conference on Technologies for e-learning and digital entertainment
A multimodal fusion framework for children's storytelling systems
Edutainment'06 Proceedings of the First international conference on Technologies for E-Learning and Digital Entertainment
Hi-index | 0.00 |
Eyesight and speech are two channels that humansnaturally use to communicate with each other.However both the eye tracking and the speechrecognition technique available today are still far fromperfect. Our goal is find how to effectively make use ofthese error-prone information from both modes, inorder to use one mode to correct errors of anothermode, overcome the immature of recognitiontechniques, resolve the ambiguity of the user'sspeaking, and improve the interaction speed. Theintegration strategies and the evaluation experimentdemonstrate that these two modalities can be usedmultimodally to improve the usability and efficiency ofuser interface, which would not be available to speech-onlyor gaze-only systems.