Elements of information theory
Elements of information theory
Machine Learning
Inducing Features of Random Fields
IEEE Transactions on Pattern Analysis and Machine Intelligence
Distributed Detection and Data Fusion
Distributed Detection and Data Fusion
IEEE Transactions on Pattern Analysis and Machine Intelligence
Transductive Inference for Text Classification using Support Vector Machines
ICML '99 Proceedings of the Sixteenth International Conference on Machine Learning
A classification paradigm for distributed vertically partitioned data
Neural Computation
Critic-driven ensemble classification
IEEE Transactions on Signal Processing
Distributed classification of Gaussian space-time sources in wireless sensor networks
IEEE Journal on Selected Areas in Communications
An Extension of Iterative Scaling for Decision and Data Aggregation in Ensemble Classification
Journal of VLSI Signal Processing Systems
Non-stationary data sequence classification using online class priors estimation
Pattern Recognition
Protein Fold Prediction Problem Using Ensemble of Classifiers
ICONIP '09 Proceedings of the 16th International Conference on Neural Information Processing: Part II
Transfer estimation of evolving class priors in data stream classification
Pattern Recognition
Hi-index | 0.00 |
We consider ensemble classification for the case where there is no common labeled training data for jointly designing the individual classifiers and the function that aggregates their decisions. This problem, which we call distributed ensemble classification, applies when individual classifiers operate (perhaps remotely) on different sensing modalities and when combining proprietary or legacy classifiers. The conventional wisdom in this case is to apply fixed rules of combination such as voting methods or rules for aggregating probabilities. Alternatively, we take a transductive approach, optimizing the combining rule for an objective function measured on the unlabeled batch of test data. We propose maximum likelihood (ML) objectives that are shown to yield well-known forms of probabilistic aggregation, albeit with iterative, expectation-maximization-based adjustment to account for mismatch between class priors used by individual classifiers and those reflected in the new data batch. These methods are extensions, for the ensemble case, of the work of Saerens, Latinne, and Decaestecker (2002). We also propose an information-theoretic method that generally outperforms the ML methods, better handles classifier redundancies, and addresses some scenarios where the ML methods are not applicable. This method also well handles the case of missing classes in the test batch. On UC Irvine benchmark data, all our methods give improvements in classification accuracy over the use of fixed rules when there is prior mismatch.