X-means: Extending K-means with Efficient Estimation of the Number of Clusters
ICML '00 Proceedings of the Seventeenth International Conference on Machine Learning
Neural Computation
Self-splitting competitive learning: a new on-line clustering paradigm
IEEE Transactions on Neural Networks
Hi-index | 0.00 |
The Self-Splitting Competitive Learning (SSCL) is a powerful algorithm that solves the difficult problems of determining the number of clusters and the sensitivity to prototype initialization in clustering. The SSCL algorithm iteratively partitions the data space into natural clusters without a prior information on the number of clusters. However, SSCL suffers from two major disadvantages: it does not have a proven convergence and the speed of learning process is slow. We propose solutions for these two problems. Firstly, we introduce a new update scheme and lead a proven convergence of Asymptotic Property Vector. Secondly, we modify the split-validity to accelerate the learning process. Experiments show these techniques make the algorithm faster than the original one.