International Journal of Man-Machine Studies - Special Issue: Knowledge Acquisition for Knowledge-based Systems. Part 5
C4.5: programs for machine learning
C4.5: programs for machine learning
A Comparative Analysis of Methods for Pruning Decision Trees
IEEE Transactions on Pattern Analysis and Machine Intelligence
Machine Learning
Using output codes to boost multiclass learning problems
ICML '97 Proceedings of the Fourteenth International Conference on Machine Learning
An Empirical Comparison of Pruning Methods for Ensemble Classifiers
IDA '01 Proceedings of the 4th International Conference on Advances in Intelligent Data Analysis
ICML '97 Proceedings of the Fourteenth International Conference on Machine Learning
Reducing multiclass to binary: a unifying approach for margin classifiers
The Journal of Machine Learning Research
Simplifying decision trees: A survey
The Knowledge Engineering Review
Solving multiclass learning problems via error-correcting output codes
Journal of Artificial Intelligence Research
Spectral coefficients and classifier correlation
MCS'03 Proceedings of the 4th international conference on Multiple classifier systems
Hi-index | 0.00 |
In this paper we consider the combination of two ensemble techniques, both capable of producing diverse binary base classifiers. Adaboost, a version of Boosting is combined with Output Coding for solving multiclass problems. Decision trees are chosen as the base classifiers, and the issue of tree pruning is addressed. Pruning produces less complex trees and sometimes leads to better generalisation. Experimental results demonstrate that pruning makes little difference in this framework. However, on average over nine benchmark datasets better accuracy is achieved by incorporating unpruned trees.