Bayesian classifiers based on kernel density estimation: Flexible classifiers
International Journal of Approximate Reasoning
A growing and pruning method for radial basis function networks
IEEE Transactions on Neural Networks
Probabilistic PCA self-organizing maps
IEEE Transactions on Neural Networks
GMM based SPECT image classification for the diagnosis of Alzheimer's disease
Applied Soft Computing
Hi-index | 35.68 |
In this paper we seek a Gaussian mixture model (GMM) of an n-variate probability density function. Usually the parameters of GMMs are determined in the original n-dimensional space by optimizing a maximum likelihood (ML) criterion. A practical deficiency of this method of fitting GMMs is its poor performance when dealing with high-dimensional data since a large sample size is needed to match the accuracy that is possible in low dimensions. We propose a method for fitting the GMM based on the projection pursuit strategy. This GMM is highly constrained and hence its ability to model structure in subspaces is enhanced, compared to a direct ML fitting of a GMM in high dimensions. Our method is closely related to recently developed independent factor analysis (IFA) mixture models. The comparisons with ML fitting of GMM in n-dimensions and IFA mixtures show that the proposed method is an attractive choice for fitting GMMs using small sizes of training sets.