IBM multimodal interactive video threading

  • Authors:
  • Jelena Tešić;Apostol Natsev;Joachim Seidl;John R. Smith

  • Affiliations:
  • IBM Watson Research Center, Hawthorne, NY;IBM Watson Research Center, Hawthorne, NY;IBM Watson Research Center, Hawthorne, NY;IBM Watson Research Center, Hawthorne, NY

  • Venue:
  • Proceedings of the 6th ACM international conference on Image and video retrieval
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this demo we present a novel approach for (a) automatic labeling and grouping of multimedia content using existing metadata and semantic concepts, and (b) interactive context driven tagging of clusters of multimedia content. Proposed system leverages existing metadata info in conjunction with automatically assigned semantic descriptors. One of the challenges of multimedia retrieval systems today is to organize and present the video data in such a way that allows the user to most efficiently navigate the rich index space. The information needs of users typically span a range of semantic concepts, associated metadata, and content similarity. We propose to jointly analyze and navigate metadata, semantic and visual space for the purpose of identifying new relationships among content, and allowing user to link the aggregated content to a complex event description. The advantages of the proposed system are realized in increased ability to target content delivery to users such as in the collaborative or multi-domain user environments.