Building Multi-Modal Relational Graphs for Multimedia Retrieval

  • Authors:
  • Ja-Ling Wu;Jyh-Ren Shieh;Ching-Yung Lin;Shun-Xuan Wang

  • Affiliations:
  • National Taiwan University, Taiwan;National Taiwan University, Taiwan;IBM T. J. Watson Research Center, USA;National Taiwan University, Taiwan

  • Venue:
  • International Journal of Multimedia Data Engineering & Management
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

The abundance of Web 2.0 social media in various media formats calls for integration that takes into account tags associated with these resources. The authors present a new approach to multi-modal media search, based on novel related-tag graphs, in which a query is a resource in one modality, such as an image, and the results are semantically similar resources in various modalities, for instance text and video. Thus the use of resource tagging enables the use of multi-modal results and multi-modal queries, a marked departure from the traditional text-based search paradigm. Tag relation graphs are built based on multi-partite networks of existing Web 2.0 social media such as Flickr and Wikipedia. These multi-partite linkage networks contributor-tag, tag-category, and tag-tag are extracted from Wikipedia to construct relational tag graphs. In fusing these networks, the authors propose incorporating contributor-category networks to model contributor's specialization; it is shown that this step significantly enhances the accuracy of the inferred relatedness of the term-semantic graphs. Experiments based on 200 TREC-5 ad-hoc topics show that the algorithms outperform existing approaches. In addition, user studies demonstrate the superiority of this visualization system and its usefulness in the real world.