An analytic distance metric for Gaussian mixture models with application in image retrieval

  • Authors:
  • G. Sfikas;C. Constantinopoulos;A. Likas;N. P. Galatsanos

  • Affiliations:
  • Department of Computer Science, University of Ioannina, Ioannina, Greece;Department of Computer Science, University of Ioannina, Ioannina, Greece;Department of Computer Science, University of Ioannina, Ioannina, Greece;Department of Computer Science, University of Ioannina, Ioannina, Greece

  • Venue:
  • ICANN'05 Proceedings of the 15th international conference on Artificial neural networks: formal models and their applications - Volume Part II
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper we propose a new distance metric for probability density functions (PDF). The main advantage of this metric is that unlike the popular Kullback-Liebler (KL) divergence it can be computed in closed form when the PDFs are modeled as Gaussian Mixtures (GM). The application in mind for this metric is histogram based image retrieval. We experimentally show that in an image retrieval scenario the proposed metric provides as good results as the KL divergence at a fraction of the computational cost. This metric is also compared to a Bhattacharyya-based distance metric that can be computed in closed form for GMs and is found to produce better results.