Voice activity detection from gaze in video mediated communication
Proceedings of the Symposium on Eye Tracking Research and Applications
Multimodal cue detection engine for orchestrated entertainment
MMM'12 Proceedings of the 18th international conference on Advances in Multimedia Modeling
The fascinate production scripting engine
MMM'12 Proceedings of the 18th international conference on Advances in Multimedia Modeling
Automatic orchestration of video streams to enhance group communication
Proceedings of the 2012 international workshop on Socially-aware multimedia
Enabling 'togetherness' in high-quality domestic video
Proceedings of the 20th ACM international conference on Multimedia
Virtual director technology for social video communication and live event broadcast production
Proceedings of the 21st ACM international conference on Multimedia
Real-time audio-visual analysis for multiperson videoconferencing
Advances in Multimedia
Hi-index | 0.00 |
In this paper we present an approach to the reasoning required to support multi-location, multi-camera group-to-group video communication, which we call orchestration. Orchestration is a kin to virtual directing: it has to ensure that each location displays the most adequate shots from all the other available sources. Its input is low-level cues extracted automatically from the AV streams. They are processed to detect higher-level events that determine the state of the communication. Directorial decisions are then inferred, reflecting social communication as well as stylistic criteria. Finally, they are transformed into camera and editing commands, directly executable by the AV infrastructure. Here, we present the architecture of the Orchestrator and sketch our rule-based approach to reasoning.