Automatic audio content analysis
MULTIMEDIA '96 Proceedings of the fourth ACM international conference on Multimedia
Evolving video skims into useful multimedia abstractions
Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
Auto-summarization of audio-video presentations
MULTIMEDIA '99 Proceedings of the seventh ACM international conference on Multimedia (Part 1)
Learning video browsing behavior and its application in the generation of video previews
MULTIMEDIA '01 Proceedings of the ninth ACM international conference on Multimedia
A utility framework for the automatic generation of audio-visual skims
Proceedings of the tenth ACM international conference on Multimedia
A user attention model for video summarization
Proceedings of the tenth ACM international conference on Multimedia
Video Content Analysis Using Multimodal Information: For Movie Content Extraction, Indexing and Representation
Contrast-based image attention analysis by using fuzzy growing
MULTIMEDIA '03 Proceedings of the eleventh ACM international conference on Multimedia
Evaluation campaigns and TRECVid
MIR '06 Proceedings of the 8th ACM international workshop on Multimedia information retrieval
Brain state decoding for rapid image retrieval
MM '09 Proceedings of the 17th ACM international conference on Multimedia
A generic framework of user attention model and its application in video summarization
IEEE Transactions on Multimedia
Using eye-tracking data for automatic film comic creation
Proceedings of the Symposium on Eye Tracking Research and Applications
Hi-index | 0.00 |
A variety of user attention models for video/audio streams have been developed for video summarization and abstraction, in order to facilitate efficient video browsing and indexing. Essentially, human brain is the end user and evaluator of multimedia content and representation, and its responses can provide meaningful guidelines for multimedia stream summarization. For example, video/audio segments that significantly activate the visual, auditory, language and working memory systems of the human brain should be considered more important than others. It should be noted that user experience studies could be useful for such evaluations, but are suboptimal in terms of their capability of accurately capturing the full-length dynamics and interactions of the brain's response. This paper presents our preliminary efforts in applying the brain imaging technique of functional magnetic resonance imaging (fMRI) to quantify and model the dynamics and interactions between multimedia streams and brain response, when the human subjects are presented with the multimedia clips, in order to develop human-centered attention models that can be used to guide and facilitate more effective and efficient multimedia summarization. Our initial results are encouraging.