A Model of Saliency-Based Visual Attention for Rapid Scene Analysis
IEEE Transactions on Pattern Analysis and Machine Intelligence
MULTIMEDIA '00 Proceedings of the eighth ACM international conference on Multimedia
Models of bottom-up and top-down visual attention
Models of bottom-up and top-down visual attention
Visual attention detection in video sequences using spatiotemporal cues
MULTIMEDIA '06 Proceedings of the 14th annual ACM international conference on Multimedia
Keyframe-Based Video Summary Using Visual Attention Clues
IEEE MultiMedia
Salient region detection using weighted feature maps based on the human visual attention model
PCM'04 Proceedings of the 5th Pacific Rim Conference on Advances in Multimedia Information Processing - Volume Part II
Two-stage hierarchical video summary extraction to match low-level user browsing preferences
IEEE Transactions on Multimedia
A generic framework of user attention model and its application in video summarization
IEEE Transactions on Multimedia
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
A novel video key-frame-extraction algorithm based on perceived motion energy model
IEEE Transactions on Circuits and Systems for Video Technology
Information theory-based shot cut/fade detection and video summarization
IEEE Transactions on Circuits and Systems for Video Technology
Efficient visual attention based framework for extracting key frames from videos
Image Communication
A novel video salient object extraction method based on visual attention
Image Communication
Saliency detection using joint spatial-color constraint and multi-scale segmentation
Journal of Visual Communication and Image Representation
Spatiotemporal saliency detection and salient region determination for H.264 videos
Journal of Visual Communication and Image Representation
Hi-index | 0.00 |
Key frame extraction is an important technique in video summarization, browsing, searching and understanding. In this paper, we propose a novel approach to extract the most attractive key frames by using a saliency-based visual attention model that bridges the gap between semantic interpretation of the video and low-level features. First, dynamic and static conspicuity maps are constructed based on motion, color and texture features. Then, by introducing suppression factor and motion priority schemes, the conspicuity maps are fused into a saliency map that includes only true attention regions to produce attention curve. Finally, after time-constraint cluster algorithm grouping frames with similar content, the frames with maximum saliency value are selected as key-frames. Experimental results demonstrate the effectiveness of our approach for video summarization by retrieving the meaningful key frames.