Weakly Supervised Learning of Visual Models and Its Application to Content-Based Retrieval

  • Authors:
  • Cordelia Schmid

  • Affiliations:
  • INRIA Rhône-Alpes, 655 av. de l'Europe, 38330 Montbonnot, France. Cordelia.Schmid@inrialpes.fr

  • Venue:
  • International Journal of Computer Vision - Special Issue on Content-Based Image Retrieval
  • Year:
  • 2004

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper presents a method for weakly supervised learning of visual models. The visual model is based on a two-layer image description: a set of “generic” descriptors and their distribution over neighbourhoods. “Generic” descriptors represent sets of similar rotational invariant feature vectors. Statistical spatial constraints describe the neighborhood structure and make our description more discriminant. The joint probability of the frequencies of “generic” descriptors over a neighbourhood is multi-modal and is represented by a set of “neighbourhood-frequency” clusters. Our image description is rotationally invariant, robust to model deformations and characterizes efficiently “appearance-based” visual structure. The selection of distinctive clusters determines model features (common to the positive and rare in the negative examples). Visual models are retrieved and localized using a probabilistic score. Experimental results for “textured” animals and faces show a very good performance for retrieval as well as localization.