Exploiting generative models in discriminative classifiers
Proceedings of the 1998 conference on Advances in neural information processing systems II
Machine Learning: Discriminative and Generative (Kluwer International Series in Engineering and Computer Science)
Improved learning of Riemannian metrics for exploratory analysis
Neural Networks - 2004 Special issue: New developments in self-organizing systems
A Nonlinear Mapping for Data Structure Analysis
IEEE Transactions on Computers
CIARP'11 Proceedings of the 16th Iberoamerican Congress conference on Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications
Information geometry on hierarchy of probability distributions
IEEE Transactions on Information Theory
Bankruptcy analysis with self-organizing maps in learning metrics
IEEE Transactions on Neural Networks
Hi-index | 0.01 |
Measures of similarity are fundamental in pattern recognition and data mining. Typically the Euclidean metric is used in this context, weighting all variables equally and therefore assuming equal relevance, which is very rare in real applications. In contrast, given an estimate of a conditional density function, the Fisher information calculated in primary data space implicitly measures the relevance of variables in a principled way by reference to auxiliary data such as class labels. This paper proposes a framework that uses a distance metric based on Fisher information to construct similarity networks that achieve a more informative and principled representation of data. The framework enables efficient retrieval of reference cases from which a weighted nearest neighbour classifier closely approximates the original density function. Importantly, the identification of nearby data points permits the retrieval of further information with potential relevance to the assessment of a new case. The practical application of the method is illustrated for six benchmark datasets.