On the induction of decision trees for multiple concept learning
On the induction of decision trees for multiple concept learning
ChiMerge: discretization of numeric attributes
AAAI'92 Proceedings of the tenth national conference on Artificial intelligence
Hi-index | 0.00 |
Most of the existing discretization methods such as k-interval discretization, equal width and equal frequency methods do not take the dependencies of decision attributes on condition attributes into account. In this paper, we propose a discretization algorithm that can keep the dependencies of the decision attribute on condition attributes, or keep the positive regions of the partition of the decision attribute. In the course of inducing classification rules from a data set, keeping these dependencies can achieve getting the set of the least condition attributes and the highest classification precision.