Attention, intentions, and the structure of discourse
Computational Linguistics
Cognitive Status and Form of Reference in Multimodal Human-Computer Interaction
Proceedings of the Seventeenth National Conference on Artificial Intelligence and Twelfth Conference on Innovative Applications of Artificial Intelligence
Multimodal Cooperative Resolution of Referential Expressions in the DenK System
CMC '98 Revised Papers from the Second International Conference on Cooperative Multimodal Communication
“Put-that-there”: Voice and gesture at the graphics interface
SIGGRAPH '80 Proceedings of the 7th annual conference on Computer graphics and interactive techniques
Hi-index | 0.00 |
The way we see the objects around us determines speech and gestures we use to refer to them. The gestures we produce structure our visual perception. The words we use have an influence on the way we see. In this manner, visual perception, language and gesture present multiple interactions between each other. The problem is global and has to be tackled as a whole in order to understand the complexity of reference phenomena and to deduce a formal model. This model may be useful for any kind of man-machine dialogue system that focuses on deep comprehension. We show how a referring act takes place into a contextual subset of objects, called ‘reference domain,' and we present the ‘multimodal reference domain' model that can be exploited in a dialogue system when interpreting.