Representation of occurrences for road vehicle traffic
Artificial Intelligence
Multi-level Particle Filter Fusion of Features and Cues for Audio-Visual Person Tracking
Multimodal Technologies for Perception of Humans
ISL Person Identification Systems in the CLEAR 2007 Evaluations
Multimodal Technologies for Perception of Humans
ICMI '08 Proceedings of the 10th international conference on Multimodal interfaces
Semantic Representation and Recognition of Continued and Recursive Human Activities
International Journal of Computer Vision
Computers in the Human Interaction Loop
Computers in the Human Interaction Loop
Understanding dynamic scenes based on human sequence evaluation
Image and Vision Computing
PERCOM '09 Proceedings of the 2009 IEEE International Conference on Pervasive Computing and Communications
Handbook of Ambient Intelligence and Smart Environments
Handbook of Ambient Intelligence and Smart Environments
Extending touch: towards interaction with large-scale surfaces
Proceedings of the ACM International Conference on Interactive Tabletops and Surfaces
Multimodal Signal Processing: Theory and applications for human-computer interaction
Multimodal Signal Processing: Theory and applications for human-computer interaction
Machine Recognition of Human Activities: A Survey
IEEE Transactions on Circuits and Systems for Video Technology
Rule-based high-level situation recognition from incomplete tracking data
RuleML'12 Proceedings of the 6th international conference on Rules on the Web: research and applications
An information retrieval approach to identifying infrequent events in surveillance video
Proceedings of the 3rd ACM conference on International conference on multimedia retrieval
Automatic unconstrained online configuration of a master-slave camera system
ICVS'13 Proceedings of the 9th international conference on Computer Vision Systems
Hi-index | 0.00 |
Most approaches to the visual perception of humans do not include high-level activity recognitition. This paper presents a system that fuses and interprets the outputs of several computer vision components as well as speech recognition to obtain a high-level understanding of the perceived scene. Our laboratory for investigating new ways of human-machine interaction and teamwork support, is equipped with an assemblage of cameras, some close-talking microphones, and a videowall as main interaction device. Here, we develop state of the art real-time computer vision systems to track and identify users, and estimate their visual focus of attention and gesture activity. We also monitor the users' speech activity in real time. This paper explains our approach to highlevel activity recognition based on these perceptual components and a temporal logic engine.