Data stability in clustering: a closer look

  • Authors:
  • Lev Reyzin

  • Affiliations:
  • School of Computer Science, Georgia Institute of Technology, Atlanta, GA

  • Venue:
  • ALT'12 Proceedings of the 23rd international conference on Algorithmic Learning Theory
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

We consider the model introduced by Bilu and Linial [12],, who study problems for which the optimal clustering does not change when distances are perturbed. They show that even when a problem is NP-hard, it is sometimes possible to obtain efficient algorithms for instances resilient to certain multiplicative perturbations, e.g. on the order of $O(\sqrt{n})$ for max-cut clustering. Awasthi et al. [6], consider center-based objectives, and Balcan and Liang [9], analyze the k-median and min-sum objectives, giving efficient algorithms for instances resilient to certain constant multiplicative perturbations. Here, we are motivated by the question of to what extent these assumptions can be relaxed while allowing for efficient algorithms. We show there is little room to improve these results by giving NP-hardness lower bounds for both the k-median and min-sum objectives. On the other hand, we show that multiplicative resilience parameters, even only on the order of Θ(1), can be so strong as to make the clustering problem trivial, and we exploit these assumptions to present a simple one-pass streaming algorithm for the k-median objective. We also consider a model of additive perturbations and give a correspondence between additive and multiplicative notions of stability. Our results provide a close examination of the consequences of assuming, even constant, stability in data.