Temporal Synchronization of Video Sequences in Theory and in Practice
WACV-MOTION '05 Proceedings of the IEEE Workshop on Motion and Video Computing (WACV/MOTION'05) - Volume 2 - Volume 02
Synchronization of multiple video recordings based on still camera flashes
MULTIMEDIA '06 Proceedings of the 14th annual ACM international conference on Multimedia
Less talk, more rock: automated organization of community-contributed collections of concert videos
Proceedings of the 18th international conference on World wide web
Virtual video camera: image-based viewpoint navigation through space and time
SIGGRAPH '09: Posters
The multiplayer: multi-perspective social video navigation
UIST '10 Adjunct proceedings of the 23nd annual ACM symposium on User interface software and technology
Automatic mashup generation from multiple-camera concert recordings
Proceedings of the international conference on Multimedia
Multimedia Tools and Applications
Making a scene: alignment of complete sets of clips based on pairwise audio match
Proceedings of the 2nd ACM International Conference on Multimedia Retrieval
MoViMash: online mobile video mashup
Proceedings of the 20th ACM international conference on Multimedia
Hi-index | 0.00 |
An increasing number of people regularly capture video in social occasions like weddings, parties and holiday trips. As a result, multiple video recordings are made from a single event providing different view angles and wider coverage. This gives an opportunity to produce a desired video summary from the event, combining the videos with the most favorable views from multiple recordings. In order to mix contents from different cameras, the recordings require very precise synchronization in time. This task is very tedious and presently done manually. We present two methods to synchronize multiple videos based on the identical audio content present in the recordings. The first method utilizes audio-classification and the synchronization between two recordings is determined by correlating the audio classes. The second method uses audio-fingerprints to represent the recorded audio. The synchronization is determined by fingerprint matches between the different recordings. The experimental results show that the audio-classification method requires recordings, at least a couple of minutes long, with large temporal overlap to determine the synchronization point. The method using audio-fingerprints requires at least 3 second long overlapping audio and resulted inperfect synchronization in all the examined cases.