Information theoretic measures for clusterings comparison: is a correction for chance necessary?

  • Authors:
  • Nguyen Xuan Vinh;Julien Epps;James Bailey

  • Affiliations:
  • The University of New South Wales, Sydney, Australia & ATP Laboratory, National ICT Australia (NICTA);The University of New South Wales, Sydney, Australia & ATP Laboratory, National ICT Australia (NICTA);The University of Melbourne, Australia & Victoria Research Laboratory, National ICT Australia

  • Venue:
  • ICML '09 Proceedings of the 26th Annual International Conference on Machine Learning
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

Information theoretic based measures form a fundamental class of similarity measures for comparing clusterings, beside the class of pair-counting based and set-matching based measures. In this paper, we discuss the necessity of correction for chance for information theoretic based measures for clusterings comparison. We observe that the baseline for such measures, i.e. average value between random partitions of a data set, does not take on a constant value, and tends to have larger variation when the ratio between the number of data points and the number of clusters is small. This effect is similar in some other non-information theoretic based measures such as the well-known Rand Index. Assuming a hypergeometric model of randomness, we derive the analytical formula for the expected mutual information value between a pair of clusterings, and then propose the adjusted version for several popular information theoretic based measures. Some examples are given to demonstrate the need and usefulness of the adjusted measures.