Optimal multimodal fusion for multimedia data analysis

  • Authors:
  • Yi Wu;Edward Y. Chang;Kevin Chen-Chuan Chang;John R. Smith

  • Affiliations:
  • University of California Santa Barbara, Santa Barbara, CA;University of California Santa Barbara, Santa Barbara, CA;University of Illinois at Urbana-Champaign, Urbana, IL;IBM T.J. Watson Research Center, Hawthorne, NY

  • Venue:
  • Proceedings of the 12th annual ACM international conference on Multimedia
  • Year:
  • 2004

Quantified Score

Hi-index 0.00

Visualization

Abstract

Considerable research has been devoted to utilizing multimodal features for better understanding multimedia data. However, two core research issues have not yet been adequately addressed. First, given a set of features extracted from multiple media sources (e.g., extracted from the visual, audio, and caption track of videos), how do we determine the best modalities? Second, once a set of modalities has been identified, how do we best fuse them to map to semantics? In this paper, we propose a two-step approach. The first step finds statistically independent modalities from raw features. In the second step, we use super-kernel fusion to determine the optimal combination of individual modalities. We carefully analyze the tradeoffs between three design factors that affect fusion performance: modality independence, curse of dimensionality, and fusion-model complexity. Through analytical and empirical studies, we demonstrate that our two-step approach, which achieves a careful balance of the three design factors, can improve class-prediction accuracy over traditional techniques.