An Empirical Comparison of Selection Measures for Decision-Tree Induction

  • Authors:
  • John Mingers

  • Affiliations:
  • School of Industrial and Business Studies, University of Warwick, Coventry CV4 7AL, U.K. BSRCD@CU.WARWICK.AC.UK

  • Venue:
  • Machine Learning
  • Year:
  • 1989

Quantified Score

Hi-index 0.00

Visualization

Abstract

One approach to induction is to develop a decision tree from a set of examples. When used with noisy rather than deterministic data, the method involves three main stages – creating a complete tree able to classify all the examples, pruning this tree to give statistical reliability, and processing the pruned tree to improve understandability. This paper is concerned with the first stage – tree creation – which relies on a measure for “goodness of split,” that is, how well the attributes discriminate between classes. Some problems encountered at this stage are missing data and multi-valued attributes. The paper considers a number of different measures and experimentally examines their behavior in four domains. The results show that the choice of measure affects the size of a tree but not its accuracy, which remains the same even when attributes are selected randomly.