Fast principal component analysis using fixed-point algorithm
Pattern Recognition Letters
Improved facial-feature detection for AVSP via unsupervised clustering and discriminant analysis
EURASIP Journal on Applied Signal Processing
A gradient procedure for determining clusters of relatively high point density
Pattern Recognition
Data clustering: 50 years beyond K-means
Pattern Recognition Letters
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
The estimation of the gradient of a density function, with applications in pattern recognition
IEEE Transactions on Information Theory
Hi-index | 0.10 |
This paper describes a new semi-supervised learning algorithm for intra-class clustering (ICC). ICC partitions each class into sub-classes in order to minimize overlap across clusters from different classes. This is achieved by allowing partitioning of a certain class to be assisted by data points from other classes in a context-dependent fashion. The result is that overlap across sub-classes (both within- and across class) is greatly reduced. ICC is particularly useful when combined with algorithms that assume that each class has a unimodal Gaussian distribution (e.g., Linear Discriminant Analysis (LDA), quadratic classifiers), an assumption that is not always true in many real-world situations. ICC can help partition non-Gaussian, multimodal distributions to overcome such a problem. In this sense, ICC works as a preprocessor. Experiments with our ICC algorithm on synthetic data sets and real-world data sets indicated that it can significantly improve the performance of LDA and quadratic classifiers. We expect our approach to be applicable to a broader class of pattern recognition problems where class-conditional densities are significantly non-Gaussian or multi-modal.