Speech Communication - Special issue on interactive voice technology for telecommunication applications (IVITA '96)
Multimodal interaction for distributed interactive simulation
Readings in intelligent user interfaces
Mutual disambiguation of recognition errors in a multimodel architecture
Proceedings of the SIGCHI conference on Human Factors in Computing Systems
Toward conversational human-computer interaction
AI Magazine
Multimodal Maps: An Agent-Based Approach
Multimodal Human-Computer Communication, Systems, Techniques, and Experiments
Speech and sketching for multimodal design
Proceedings of the 9th international conference on Intelligent user interfaces
Gemini: a natural language system for spoken-language understanding
ACL '93 Proceedings of the 31st annual meeting on Association for Computational Linguistics
Deixis and conjunction in multimodal systems
COLING '00 Proceedings of the 18th conference on Computational linguistics - Volume 1
Finite-state multimodal parsing and understanding
COLING '00 Proceedings of the 18th conference on Computational linguistics - Volume 1
Finite-state multimodal integration and understanding
Natural Language Engineering
MATCH: an architecture for multimodal dialogue systems
ACL '02 Proceedings of the 40th Annual Meeting on Association for Computational Linguistics
Understanding spontaneous speech: the Phoenix system
ICASSP '91 Proceedings of the Acoustics, Speech, and Signal Processing, 1991. ICASSP-91., 1991 International Conference
Multimodal interactive maps: designing for human performance
Human-Computer Interaction
Robust understanding in multimodal interfaces
Computational Linguistics
Hi-index | 0.00 |
With the explosive growth in mobile computing and communication over the past few years, it is possible to access almost any information from virtually anywhere. However, the efficiency and effectiveness of this interaction is severely limited by the inherent characteristics of mobile devices, including small screen size and the lack of a viable keyboard or mouse. This paper concerns the use of multimodal language processing techniques to enable interfaces combining speech and gesture input that overcome these limitations. Specifically we focus on robust processing of pen gesture inputs in a local search application and demonstrate that edit-based techniques that have proven effective in spoken language processing can also be used to overcome unexpected or errorful gesture input. We also examine the use of a bottom-up gesture aggregation technique to improve the coverage of multimodal understanding.