IMPACT: an interactive natural-motion-picture dedicated multimedia authoring system
CHI '91 Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
Video parsing, retrieval and browsing: an integrated and content-based solution
Proceedings of the third ACM international conference on Multimedia
Video Rewrite: driving visual speech with audio
Proceedings of the 24th annual conference on Computer graphics and interactive techniques
Synthesizing realistic facial expressions from photographs
Proceedings of the 25th annual conference on Computer graphics and interactive techniques
Warping and morphing of graphical objects
Warping and morphing of graphical objects
An optimal algorithm for approximate nearest neighbor searching fixed dimensions
Journal of the ACM (JACM)
A morphable model for the synthesis of 3D faces
Proceedings of the 26th annual conference on Computer graphics and interactive techniques
Proceedings of the 27th annual conference on Computer graphics and interactive techniques
A semi-automatic approach to home video editing
UIST '00 Proceedings of the 13th annual ACM symposium on User interface software and technology
Controlled animation of video sprites
Proceedings of the 2002 ACM SIGGRAPH/Eurographics symposium on Computer animation
Simplifying video editing using metadata
DIS '02 Proceedings of the 4th conference on Designing interactive systems: processes, practices, methods, and techniques
Nonlinear manifold learning for visual speech recognition
ICCV '95 Proceedings of the Fifth International Conference on Computer Vision
Graphcut textures: image and video synthesis using graph cuts
ACM SIGGRAPH 2003 Papers
Spectral Grouping Using the Nyström Method
IEEE Transactions on Pattern Analysis and Machine Intelligence
Spacetime faces: high resolution capture for modeling and animation
ACM SIGGRAPH 2004 Papers
Histograms of Oriented Gradients for Human Detection
CVPR '05 Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Volume 1 - Volume 01
ACM SIGGRAPH 2005 Papers
Video abstraction: A systematic review and classification
ACM Transactions on Multimedia Computing, Communications, and Applications (TOMCCAP)
Space-Time Completion of Video
IEEE Transactions on Pattern Analysis and Machine Intelligence
Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
Video browsing by direct manipulation
Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
DRAGON: a direct manipulation interface for frame-accurate in-scene video navigation
Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
Video object annotation, navigation, and composition
Proceedings of the 21st annual ACM symposium on User interface software and technology
Moving gradients: a path-based method for plausible image interpolation
ACM SIGGRAPH 2009 papers
ECCV'10 Proceedings of the 11th European conference on Computer vision: Part I
BoostMap: a method for efficient approximate similarity rankings
CVPR'04 Proceedings of the 2004 IEEE computer society conference on Computer vision and pattern recognition
ACM SIGGRAPH 2011 papers
Proceedings of the 2011 SIGGRAPH Asia Conference
Content-based tools for editing audio stories
Proceedings of the 26th annual ACM symposium on User interface software and technology
DemoCut: generating concise instructional videos for physical demonstrations
Proceedings of the 26th annual ACM symposium on User interface software and technology
Hi-index | 0.00 |
We present a set of tools designed to help editors place cuts and create transitions in interview video. To help place cuts, our interface links a text transcript of the video to the corresponding locations in the raw footage. It also visualizes the suitability of cut locations by analyzing the audio/visual features of the raw footage to find frames where the speaker is relatively quiet and still. With these tools editors can directly highlight segments of text, check if the endpoints are suitable cut locations and if so, simply delete the text to make the edit. For each cut our system generates visible (e.g. jump-cut, fade, etc.) and seamless, hidden transitions. We present a hierarchical, graph-based algorithm for efficiently generating hidden transitions that considers visual features specific to interview footage. We also describe a new data-driven technique for setting the timing of the hidden transition. Finally, our tools offer a one click method for seamlessly removing 'ums' and repeated words as well as inserting natural-looking pauses to emphasize semantic content. We apply our tools to edit a variety of interviews and also show how they can be used to quickly compose multiple takes of an actor narrating a story.