An Evaluation of Intrinsic Dimensionality Estimators
IEEE Transactions on Pattern Analysis and Machine Intelligence
Intrinsic Dimensionality Estimation With Optimally Topology Preserving Maps
IEEE Transactions on Pattern Analysis and Machine Intelligence
Nonlinear component analysis as a kernel eigenvalue problem
Neural Computation
EM algorithms for PCA and SPCA
NIPS '97 Proceedings of the 1997 conference on Advances in neural information processing systems 10
Nonlinear Modeling of Scattered Multivariate Data and Its Application to Shape Change
IEEE Transactions on Pattern Analysis and Machine Intelligence
Mixtures of probabilistic principal component analyzers
Neural Computation
Proceedings of the 1998 conference on Advances in neural information processing systems II
Estimating the Intrinsic Dimension of Data with a Fractal-Based Method
IEEE Transactions on Pattern Analysis and Machine Intelligence
Products of Gaussians and probabilistic minor component analysis
Neural Computation
Selection of Generative Models in Classification
IEEE Transactions on Pattern Analysis and Machine Intelligence
An Algorithm for Finding Intrinsic Dimensionality of Data
IEEE Transactions on Computers
Bayesian Regularization for Normal Mixture Estimation and Model-Based Clustering
Journal of Classification
Intrinsic dimension estimation of manifolds by incising balls
Pattern Recognition
An Intrinsic Dimensionality Estimator from Near-Neighbor Information
IEEE Transactions on Pattern Analysis and Machine Intelligence
Inferring the eigenvalues of covariance matrices from limited,noisy data
IEEE Transactions on Signal Processing
Parsimonious Mahalanobis kernel for the classification of high dimensional data
Pattern Recognition
Model-based clustering of high-dimensional data: A review
Computational Statistics & Data Analysis
Hi-index | 0.10 |
A central issue in dimension reduction is choosing a sensible number of dimensions to be retained. This work demonstrates the surprising result of the asymptotic consistency of the maximum likelihood criterion for determining the intrinsic dimension of a dataset in an isotropic version of probabilistic principal component analysis (PPCA). Numerical experiments on simulated and real datasets show that the maximum likelihood criterion can actually be used in practice and outperforms existing intrinsic dimension selection criteria in various situations. This paper exhibits and outlines the limits of the maximum likelihood criterion. It leads to recommend the use of the AIC criterion in specific situations. A useful application of this work would be the automatic selection of intrinsic dimensions in mixtures of isotropic PPCA for classification.