A Multi-Modal Approach to Story Segmentation for News Video

  • Authors:
  • Lekha Chaisorn;Tat-Seng Chua;Chin-Hui Lee

  • Affiliations:
  • The School of Computing, National University of Singapore, 117543, Singapore Lekhacha@comp.nus.edu.sg;The School of Computing, National University of Singapore, 117543, Singapore chuats@comp.nus.edu.sg;The School of Computing, National University of Singapore, 117543, Singapore chl@comp.nus.edu.sg

  • Venue:
  • World Wide Web
  • Year:
  • 2003

Quantified Score

Hi-index 0.01

Visualization

Abstract

This research proposes a two-level, multi-modal framework to perform the segmentation and classification of news video into single-story semantic units. The video is analyzed at the shot and story unit (or scene) levels using a variety of features and techniques. At the shot level, we employ Decision Trees technique to classify the shots into one of 13 predefined categories or mid-level features. At the scene/story level, we perform the HMM (Hidden Markov Models) analysis to locate story boundaries. Our initial results indicate that we could achieve a high accuracy of over 95% for shot classification, and over 89% in F1 measure on scene/story boundary detection. Detailed analysis reveals that HMM is effective in identifying dominant features, which helps in locating story boundaries. Our eventual goal is to support the retrieval of news video at story unit level, together with associated texts retrieved from related news sites on the web.