An empirical comparison of four initialization methods for the K-Means algorithm
Pattern Recognition Letters
Pattern Classification (2nd Edition)
Pattern Classification (2nd Edition)
Pattern Recognition Algorithms for Data Mining: Scalability, Knowledge Discovery, and Soft Granular Computing
Clustering For Data Mining: A Data Recovery Approach (Chapman & Hall/Crc Computer Science)
Clustering For Data Mining: A Data Recovery Approach (Chapman & Hall/Crc Computer Science)
Initializing K-means Batch Clustering: A Critical Evaluation of Several Techniques
Journal of Classification
k-means++: the advantages of careful seeding
SODA '07 Proceedings of the eighteenth annual ACM-SIAM symposium on Discrete algorithms
Constrained Intelligent K-Means: Improving Results with Limited Previous Knowledge.
ADVCOMP '08 Proceedings of the 2008 The Second International Conference on Advanced Engineering Computing and Applications in Sciences
Clustering
Data clustering: 50 years beyond K-means
Pattern Recognition Letters
On initializations for the minkowski weighted k-means
IDA'12 Proceedings of the 11th international conference on Advances in Intelligent Data Analysis
Hi-index | 0.00 |
This paper presents an analysis of the number of iterations K-Means takes to converge under different initializations. We have experimented with seven initialization algorithms in a total of 37 real and synthetic datasets. We have found that hierarchical-based initializations tend to be most effective at reducing the number of iterations, especially a divisive algorithm using the Ward criterion when applied to real datasets.