C4.5: programs for machine learning
C4.5: programs for machine learning
Decision Tree Induction Based on Efficient Tree Restructuring
Machine Learning
Machine Learning
Data Mining: Practical Machine Learning Tools and Techniques, Second Edition (Morgan Kaufmann Series in Data Management Systems)
Anytime Learning of Decision Trees
The Journal of Machine Learning Research
A compression algorithm for large arity extensional constraints
CP'07 Proceedings of the 13th international conference on Principles and practice of constraint programming
Itemset mining: A constraint programming perspective
Artificial Intelligence
Hi-index | 0.00 |
Decision tree induction techniques attempt to find small trees that fit a training set of data. This preference for smaller trees, which provides a learning bias, is often justified as being consistent with the principle of Occam's Razor. Informally, this principle states that one should prefer the simpler hypothesis. In this paper we take this principle to the extreme. Specifically, we formulate decision tree induction as a combinatorial optimisation problem in which the objective is to minimise the number of nodes in the tree. We study alternative formulations based on satisfiability, constraint programming, and hybrids with integer linear programming. We empirically compare our approaches against standard induction algorithms, showing that the decision trees we obtain can sometimes be less than half the size of those found by other greedy methods. Furthermore, our decision trees are competitive in terms of accuracy on a variety of well-known benchmarks, often being the most accurate. Even when post-pruning of greedy trees is used, our constraint-based approach is never dominated by any of the existing techniques.