Statistical Inference for Spatial Processes
Statistical Inference for Spatial Processes
Classification of binary vectors by stochastic complexity
Journal of Multivariate Analysis
Pattern Recognition and Neural Networks
Pattern Recognition and Neural Networks
Pattern Classification (2nd Edition)
Pattern Classification (2nd Edition)
Monte Carlo Statistical Methods (Springer Texts in Statistics)
Monte Carlo Statistical Methods (Springer Texts in Statistics)
Semi-supervised clustering: probabilistic models, algorithms and experiments
Semi-supervised clustering: probabilistic models, algorithms and experiments
Bayesian model learning based on a parallel MCMC strategy
Statistics and Computing
Approximating discrete probability distributions with dependence trees
IEEE Transactions on Information Theory
Hi-index | 0.00 |
A general inductive Bayesian classification framework is considered using a simultaneous predictive distribution for test items. We introduce a principle of generative supervised and semi-supervised classification based on marginalizing the joint posterior distribution of labels for all test items. The simultaneous and marginalized classifiers arise under different loss functions, while both acknowledge jointly all uncertainty about the labels of test items and the generating probability measures of the classes. We illustrate for data from multiple finite alphabets that such classifiers achieve higher correct classification rates than a standard marginal predictive classifier which labels all test items independently, when training data are sparse. In the supervised case for multiple finite alphabets the simultaneous and the marginal classifiers are proven to become equal under generalized exchangeability when the amount of training data increases. Hence, the marginal classifier can be interpreted as an asymptotic approximation to the simultaneous classifier for finite sets of training data. It is also shown that such convergence is not guaranteed in the semi-supervised setting, where the marginal classifier does not provide a consistent approximation.