Semantic-Context-Based augmented descriptor for image feature matching

  • Authors:
  • Samir Khoualed;Thierry Chateau;Umberto Castellani

  • Affiliations:
  • Institut Pascal, CNRS/University of Blaise Pascal, Clermont-Ferrand, France;Institut Pascal, CNRS/University of Blaise Pascal, Clermont-Ferrand, France;VIPS, University of Verona, Verona, Italy

  • Venue:
  • ACCV'12 Proceedings of the 11th Asian conference on Computer Vision - Volume Part II
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper proposes an augmented version of local feature that enhances the discriminative power of the feature without affecting its invariance to image deformations. The idea is about learning local features, aiming to estimate its semantic, which is then exploited in conjunction with the bag of words paradigm to build an augmented feature descriptor. Basically, any local descriptor can be casted in the proposed context, and thus the approach can be easy generalized to fit in with any local approach. The semantic-context signature is a 2D histogram which accumulates the spatial distribution of the visual words around each local feature. The obtained semantic-context component is concatenated with the local feature to generate our proposed feature descriptor. This is expected to handle ambiguities occurring in images with multiple similar motifs and depicting slight complicated non-affine distortions, outliers, and detector errors. The approach is evaluated for two data sets. The first one is intentionally selected with images containing multiple similar regions and depicting slight non-affine distortions. The second is the standard data set of Mikolajczyk. The evaluation results showed our approach performs significantly better than expected results as well as in comparison with other methods.