A new approach to cross-modal multimedia retrieval

  • Authors:
  • Nikhil Rasiwasia;Jose Costa Pereira;Emanuele Coviello;Gabriel Doyle;Gert R.G. Lanckriet;Roger Levy;Nuno Vasconcelos

  • Affiliations:
  • University of California, San Diego, San Diego, USA;University of California, San Diego, San Diego, USA;University of California, San Diego, San Diego, USA;University of California, San Diego, San Diego, USA;University of California, San Diego, San Diego, USA;University of California, San Diego, San Diego, USA;University of California, San Diego, San Diego, USA

  • Venue:
  • Proceedings of the international conference on Multimedia
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

The problem of joint modeling the text and image components of multimedia documents is studied. The text component is represented as a sample from a hidden topic model, learned with latent Dirichlet allocation, and images are represented as bags of visual (SIFT) features. Two hypotheses are investigated: that 1) there is a benefit to explicitly modeling correlations between the two components, and 2) this modeling is more effective in feature spaces with higher levels of abstraction. Correlations between the two components are learned with canonical correlation analysis. Abstraction is achieved by representing text and images at a more general, semantic level. The two hypotheses are studied in the context of the task of cross-modal document retrieval. This includes retrieving the text that most closely matches a query image, or retrieving the images that most closely match a query text. It is shown that accounting for cross-modal correlations and semantic abstraction both improve retrieval accuracy. The cross-modal model is also shown to outperform state-of-the-art image retrieval systems on a unimodal retrieval task.