Tracking Focus of Attention in Meetings
ICMI '02 Proceedings of the 4th IEEE International Conference on Multimodal Interfaces
Tracking the multi person wandering visual focus of attention
Proceedings of the 8th international conference on Multimodal interfaces
Audio-visual multi-person tracking and identification for smart environments
Proceedings of the 15th international conference on Multimedia
Face recognition in smart rooms
MLMI'07 Proceedings of the 4th international conference on Machine learning for multimodal interaction
Visual activity context for focus of attention estimation in dynamic meetings
ICME'09 Proceedings of the 2009 IEEE international conference on Multimedia and Expo
On the contextual analysis of agreement scores
Multimodal corpora
Putting the pieces together: multimodal analysis of social attention in meetings
Proceedings of the international conference on Multimedia
Towards high-level human activity recognition through computer vision and temporal logic
KI'10 Proceedings of the 33rd annual German conference on Advances in artificial intelligence
International Conference on Multimodal Interfaces and the Workshop on Machine Learning for Multimodal Interaction
Visual focus of attention recognition in the ambient kitchen
ACCV'09 Proceedings of the 9th Asian conference on Computer Vision - Volume Part III
Investigating the midline effect for visual focus of attention recognition
Proceedings of the 14th ACM international conference on Multimodal interaction
Recognizing the visual focus of attention for human robot interaction
HBU'12 Proceedings of the Third international conference on Human Behavior Understanding
Proceedings of the 15th ACM on International conference on multimodal interaction
Hi-index | 0.00 |
This paper presents our work on recognizing the visual focus of attention during dynamic meeting scenarios. We collected a new dataset of meetings, in which acting participants were to follow a predefined script of events, to enforce focus shifts of the remaining, unaware meeting members. Including the whole room, all in all, a total of 35 potential focus targets were annotated, of which some were moved or introduced spontaneously during the meeting. On this dynamic dataset, we present a new approach to deduce the visual focus by means of head orientation as a first clue and show, that our system recognizes the correct visual target in over 57% of all frames, compared to 47% when mapping head pose to the first-best intersecting focus target directly.