Modeling, classifying and annotating weakly annotated images using Bayesian network

  • Authors:
  • Sabine Barrat;Salvatore Tabbone

  • Affiliations:
  • LORIA-UMR 7503, BP 239, 54506 Vanduvre-les-Nancy Cedex, France;LORIA-UMR 7503, BP 239, 54506 Vanduvre-les-Nancy Cedex, France

  • Venue:
  • Journal of Visual Communication and Image Representation
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we propose a probabilistic graphical model to represent weakly annotated images. We consider an image as weakly annotated if the number of keywords defined for it is less than the maximum number defined in the ground truth. This model is used to classify images and automatically extend existing annotations to new images by taking into account semantic relations between keywords. The proposed method has been evaluated in visual-textual classification and automatic annotation of images. The visual-textual classification is performed by using both visual and textual information. The experimental results, obtained from a database of more than 30,000 images, show an improvement by 50.5% in terms of recognition rate against only visual information classification. Taking into account semantic relations between keywords improves the recognition rate by 10.5%. Moreover, the proposed model can be used to extend existing annotations to weakly annotated images, by computing distributions of missing keywords. Semantic relations improve the mean rate of good annotations by 6.9%. Finally, the proposed method is competitive with a state-of-art model.