Efficient similarity search for covariance matrices via the Jensen-Bregman LogDet Divergence

  • Authors:
  • Anoop Cherian;Suvrit Sra;Arindam Banerjee;Nikolaos Papanikolopoulos

  • Affiliations:
  • Dept. of CS & Engg., University of Minnesota, USA;MPI for Intelligent Systems, Tübingen, Germany;Dept. of CS & Engg., University of Minnesota, USA;Dept. of CS & Engg., University of Minnesota, USA

  • Venue:
  • ICCV '11 Proceedings of the 2011 International Conference on Computer Vision
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Covariance matrices provide compact, informative feature descriptors for use in several computer vision applications, such as people-appearance tracking, diffusion-tensor imaging, activity recognition, among others. A key task in many of these applications is to compare different covariance matrices using a (dis)similarity function. A natural choice here is the Riemannian metric corresponding to the manifold inhabited by covariance matrices. But computations involving this metric are expensive, especially for large matrices and even more so, in gradient-based algorithms. To alleviate these difficulties, we advocate a novel dissimilarity measure for covariance matrices: the Jensen-Bregman LogDet Divergence. This divergence enjoys several useful theoretical properties, but its greatest benefits are: (i) lower computational costs (compared to standard approaches); and (ii) amenability for use in nearest-neighbor retrieval. We show numerous experiments to substantiate these claims.