Integrating simultaneous input from speech, gaze, and hand gestures
Intelligent multimedia interfaces
Experiences with an interactive museum tour-guide robot
Artificial Intelligence - Special issue on applications of artificial intelligence
Evaluation of eye gaze interaction
Proceedings of the SIGCHI conference on Human Factors in Computing Systems
Communications of the ACM
Gaze and Speech in Attentive User Interfaces
ICMI '00 Proceedings of the Third International Conference on Advances in Multimodal Interfaces
A real-time head nod and shake detector
Proceedings of the 2001 workshop on Perceptive user interfaces
BT Technology Journal
Speech-augmented eye gaze interaction with small closely spaced targets
Proceedings of the 2006 symposium on Eye tracking research & applications
Networked reminiscence therapy for individuals with dementia by using photo and video sharing
Proceedings of the 8th international ACM SIGACCESS conference on Computers and accessibility
Multimodalcues for addressee-hood in triadic communication with a human information retrieval agent
Proceedings of the 9th international conference on Multimodal interfaces
Integrated speech and gaze control for realistic desktop environments
Proceedings of the 2008 symposium on Eye tracking research & applications
Context-Aware Computing Applications
WMCSA '94 Proceedings of the 1994 First Workshop on Mobile Computing Systems and Applications
Proceedings of the 3rd ACM International Workshop on Context-Awareness for Self-Managing Systems
Hi-index | 0.00 |
This paper proposes a videophone conversation support system by the behaviors of a companion robot and the switching of camera images in coordination with the user's conversational attitude toward the communication. In order to maintain a conversation and to achieve comfortable communication, it is necessary to understand a user's conversational states, which are whether the user is talking (taking the initiative) and whether the user is concentrating on the conversation. First, a) the system estimates the user's conversational state by a machine learning method. Next, b-1) the robot appropriately expresses its active listening behaviors, such as nodding and gaze turns, to compensate for the listener's attitude when she/he is not really listening to another user's speech, b-2) the robot shows communication-evoking behaviors (topic provision) to compensate for the lack of a topic, and b-3) the system switches the camera images to create an illusion of eye-contact corresponding to the current context of the user's attitude. From empirical studies, a detailed experiment, and a demonstration experiment, i) both the robot's active listening behaviors and the switching of the camera image compensate for the other person's attitude, ii) the topic provision function is effective for awkward silences, and iii) elderly people prefer long intervals between the robot's behaviors.