On the quest for easy-to-understand splitting rules

  • Authors:
  • Fernando Berzal;Juan-Carlos Cubero;Fernando Cuenca;María J. Martín-Bautista

  • Affiliations:
  • Department of Computer Science and Artificial Intelligence, E.T.S. Ingenieria Informática, University of Granada, Granada 18071, Spain;Department of Computer Science and Artificial Intelligence, E.T.S. Ingenieria Informática, University of Granada, Granada 18071, Spain;Xfera, Madrid, Spain;Department of Computer Science and Artificial Intelligence, E.T.S. Ingenieria Informática, University of Granada, Granada 18071, Spain

  • Venue:
  • Data & Knowledge Engineering
  • Year:
  • 2003

Quantified Score

Hi-index 0.00

Visualization

Abstract

Decision trees are probably the most popular and commonly used classification model. They are built recursively following a top-down approach (from general concepts to particular examples) by repeated splits of the training dataset. The chosen splitting criterion may affect the accuracy of the classifier, but not significantly. In fact, none of the proposed splitting criteria in the literature has proved to be universally better than the rest. Although they all yield similar results, their complexity varies significantly, and they are not always suitable for multi-way decision trees. Here we propose two new splitting rules which obtain similar results to other well-known criteria when used to build multi-way decision trees, while their simplicity makes them ideal for non-expert users.