Story segmentation in news videos using visual and text cues

  • Authors:
  • Yun Zhai;Alper Yilmaz;Mubarak Shah

  • Affiliations:
  • School of Computer Science, University of Central Florida, Orlando, Florida;School of Computer Science, University of Central Florida, Orlando, Florida;School of Computer Science, University of Central Florida, Orlando, Florida

  • Venue:
  • CIVR'05 Proceedings of the 4th international conference on Image and Video Retrieval
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we present a framework for segmenting the news programs into different story topics. The proposed method utilizes both visual and text information of the video. We represent the news video by a Shot Connectivity Graph (SCG), where the nodes in the graph represent the shots in the video, and the edges between nodes represent the transitions between shots. The cycles in the graph correspond to the story segments in the news program. We first detect the cycles in the graph by finding the anchor persons in the video. This provides us with the coarse segmentation of the news video. The initial segmentation is later refined by the detections of the weather and sporting news, and the merging of similar stories. For the weather detection, the global color information of the images and the motion of the shots are considered. We have used the text obtained from automatic speech recognition (ASR) for detecting the potential sporting shots to form the sport stories. Adjacent stories with similar semantic meanings are further merged based on the visual and text similarities. The proposed framework has been tested on a widely used data set provided by NIST, which contains the ground truth of the story boundaries, and competitive evaluation results have been obtained.