Discriminative model fusion for semantic concept detection and annotation in video

  • Authors:
  • G. Iyengar;H. J. Nock

  • Affiliations:
  • IBM TJ Watson Research Center, NY;IBM TJ Watson Research Center, NY

  • Venue:
  • MULTIMEDIA '03 Proceedings of the eleventh ACM international conference on Multimedia
  • Year:
  • 2003

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper we describe a general information fusion algorithm that can be used to incorporate multimodal cues in building user-defined semantic concept models. We compare this technique with a Bayesian Network-based approach on a semantic concept detection task. Results indicate that this technique yields superior performance. We demonstrate this approach further by building classifiers of arbitrary concepts in a score space defined by a pre-deployed set of multimodal concepts. Results show annotation for user-defined concepts both in and outside the pre-deployed set is competitive with our best video-only models on the TREC Video 2002 corpus.