A gaze-responsive self-disclosing display
CHI '90 Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
Communications of the ACM - Special issue on graphical user interfaces
Practical programming in Tcl and Tk
Practical programming in Tcl and Tk
PVM: Parallel virtual machine: a users' guide and tutorial for networked parallel computing
PVM: Parallel virtual machine: a users' guide and tutorial for networked parallel computing
ACM Computing Surveys (CSUR)
Eye tracking in advanced interface design
Virtual environments and advanced interface design
New technological windows into mind: there is more in eyes and brains for human-computer interaction
Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
Developing visual applications: XIL: an imaging foundation library
Developing visual applications: XIL: an imaging foundation library
Neural Network Perception for Mobile Robot Guidance
Neural Network Perception for Mobile Robot Guidance
Being Digital
Visual Attention and Cognition
Visual Attention and Cognition
Intelligent User Interfaces
Expert Systems: Principles and Programming
Expert Systems: Principles and Programming
LAFTER: Lips and Face Real-Time Tracker
CVPR '97 Proceedings of the 1997 Conference on Computer Vision and Pattern Recognition (CVPR '97)
Non-Intrusive Gaze Tracking Using Artificial Neural Networks
Non-Intrusive Gaze Tracking Using Artificial Neural Networks
Hi-index | 0.00 |
Research in human/computer interaction has mainly focused on natural language, text, speech and vision primarily in isolation. Recently there have been a number of research projects that have concentrated on the integration of such modalities using intelligent reasoners. The rationale is that many inherent ambiguities in single modes of communication can be resolved if extra information is available.This paper describes an intelligent multi-modal system called the Smart Work Manager. The main characteristics of the Smart Work Manager are that it can process speech, text, face images, gaze information and simulated gestures using the mouse as input modalities, and its output is in the form of speech, text or graphics. The main components of the system are the reasoner, a speech system, a vision system, an integration platform and the application interface. The overall architecture of the system will be described together with the integration platform and the components of the system which include a non-intrusive neural network based gaze-tracking system. The paper concludes with a discussion on the applicability of such systems to intelligent human/computer interaction and lessons learnt in terms of reliability and efficiency.