Human-computer interaction: input devices
ACM Computing Surveys (CSUR)
The FERET Evaluation Methodology for Face-Recognition Algorithms
IEEE Transactions on Pattern Analysis and Machine Intelligence
ACM Computing Surveys (CSUR)
Distortion Invariant Object Recognition in the Dynamic Link Architecture
IEEE Transactions on Computers
Computer vision in the interface
Communications of the ACM - Multimodal interfaces that flex, adapt, and persist
Robust Real-Time Face Detection
International Journal of Computer Vision
From conversational tooltips to grounded discourse: head poseTracking in interactive dialog systems
Proceedings of the 6th international conference on Multimodal interfaces
Object Recognition with Features Inspired by Visual Cortex
CVPR '05 Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Volume 2 - Volume 02
Recognizing gaze aversion gestures in embodied conversational discourse
Proceedings of the 8th international conference on Multimodal interfaces
Object Class Recognition and Localization Using Sparse Features with Limited Receptive Fields
International Journal of Computer Vision
Social signals, their function, and automatic analysis: a survey
ICMI '08 Proceedings of the 10th international conference on Multimodal interfaces
Head Pose Estimation in Computer Vision: A Survey
IEEE Transactions on Pattern Analysis and Machine Intelligence
Computer Vision: Algorithms and Applications
Computer Vision: Algorithms and Applications
Multiple face detection at different resolutions for perceptual user interfaces
IbPRIA'05 Proceedings of the Second Iberian conference on Pattern Recognition and Image Analysis - Volume Part I
How low level observations can help to reveal the user's state in HCI
ACII'11 Proceedings of the 4th international conference on Affective computing and intelligent interaction - Volume Part II
A companion technology for cognitive technical systems
COST'11 Proceedings of the 2011 international conference on Cognitive Behavioural Systems
Hi-index | 0.00 |
The automatic processing and estimation of view direction and head pose in interactive scenarios is an actively investigated research topic in the development of advanced human-computer or human-robot interfaces. Still, current state of the art approaches often make rigid assumptions concerning the scene illumination and viewing distance in order to achieve stable results. In addition, there is a lack of rigorous evaluation criteria to compare different computational vision approaches and to judge their flexibility. In this work, we make a step towards the employment of robust computational vision mechanisms to estimate the actor's head pose and thus the direction of his focus of attention. We propose a domain specific mechanism based on learning to estimate stereo correspondences of image pairs. Furthermore, in order to facilitate the evaluation of computational vision results, we present a data generation framework capable of image synthesis under controlled pose conditions using an arbitrary camera setup with a free number of cameras. We show some computational results of our proposed mechanism as well as an evaluation based on the available reference data.