Web and Personal Image Annotation by Mining Label Correlation With Relaxed Visual Graph Embedding

  • Authors:
  • Yi Yang;Fei Wu;Feiping Nie;Heng Tao Shen;Yueting Zhuang;Alexander G. Hauptmann

  • Affiliations:
  • School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA;College of Computer Science, Zhejiang University, Hangzhou, China;Department of Computer Science and Engineering, University of Texas, Arlington, TX, USA;School of Information Technology and Electrical Engineering, The University of Queensland, Brisbane, Australia;College of Computer Science, Zhejiang University, Hangzhou, China;School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA

  • Venue:
  • IEEE Transactions on Image Processing
  • Year:
  • 2012

Quantified Score

Hi-index 0.01

Visualization

Abstract

The number of digital images rapidly increases, and it becomes an important challenge to organize these resources effectively. As a way to facilitate image categorization and retrieval, automatic image annotation has received much research attention. Considering that there are a great number of unlabeled images available, it is beneficial to develop an effective mechanism to leverage unlabeled images for large-scale image annotation. Meanwhile, a single image is usually associated with multiple labels, which are inherently correlated to each other. A straightforward method of image annotation is to decompose the problem into multiple independent single-label problems, but this ignores the underlying correlations among different labels. In this paper, we propose a new inductive algorithm for image annotation by integrating label correlation mining and visual similarity mining into a joint framework. We first construct a graph model according to image visual features. A multilabel classifier is then trained by simultaneously uncovering the shared structure common to different labels and the visual graph embedded label prediction matrix for image annotation. We show that the globally optimal solution of the proposed framework can be obtained by performing generalized eigen-decomposition. We apply the proposed framework to both web image annotation and personal album labeling using the NUS-WIDE, MSRA MM 2.0, and Kodak image data sets, and the AUC evaluation metric. Extensive experiments on large-scale image databases collected from the web and personal album show that the proposed algorithm is capable of utilizing both labeled and unlabeled data for image annotation and outperforms other algorithms.