Using evidences based on natural language to drive the process of fusing multimodal sources

  • Authors:
  • Sergio Navarro;Fernando Llopis;Rafael Muñoz

  • Affiliations:
  • Natural Language Processing and Information Systems Group, University of Alicante, Spain;Natural Language Processing and Information Systems Group, University of Alicante, Spain;Natural Language Processing and Information Systems Group, University of Alicante, Spain

  • Venue:
  • NLDB'09 Proceedings of the 14th international conference on Applications of Natural Language to Information Systems
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper focuses on the proposal and evaluation of two multimodal fusion techniques in the field of Visual Information Retrieval (VIR). These proposals are based on two widely used fusion strategies in the VIR area, the multimodal blind relevance feedback and the multimodal re-ranking strategy. Unlike the existent techniques, our alternative proposals are guided by the evidence found in the natural language annotations related to the images. The results achieved by our runs in two different ImageCLEF tasks, 3rd place in the Wikipedia task [1] and 4th place within all the automatic runs in the photo task [2], jointly with the results obtained in later experiments presented in this paper show us that the use of conceptual information associated with an image can improve significantly the performance of the original multimodal fusion techniques used.