Clustering for text and image-based photo retrieval at CLEF 2009

  • Authors:
  • Qian Zhu;Diana Inkpen

  • Affiliations:
  • School of Information Technology and Engineering, University of Ottawa;School of Information Technology and Engineering, University of Ottawa

  • Venue:
  • CLEF'09 Proceedings of the 10th international conference on Cross-language evaluation forum: multimedia experiments
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

For this year's Image CLEF Photo Retrieval task, we investigated the effectiveness of 1) image content-based retrieval, 2) text-based retrieval, and 3) integrated text and image retrieval. We investigated whether the clustering of results can increase diversity by returning as many different clusters of images in the results as possible. Our image system used the FIRE engine to extract image features such as color, texture, and shape from a data collection consisting of about half a million images. The text-retrieval backend used Lucene to extract texts from image annotations, title, and cluster tags. Our results revealed that among the three image features, color yields the highest retrieval precision, followed by shape, then texture. A combination of color extraction with text retrieval increased precision, but only to a certain extent. Clustering also improved diversity, only in our text-based clustering runs.