Handling numeric attributes when comparing Bayesian network classifiers: does the discretization method matter?

  • Authors:
  • M. Julia Flores;José A. Gámez;Ana M. Martínez;José M. Puerta

  • Affiliations:
  • Computer Systems Department, Intelligent Systems & Data Mining--SIMD, I3A, University of Castilla-La Mancha, Albacete, Spain;Computer Systems Department, Intelligent Systems & Data Mining--SIMD, I3A, University of Castilla-La Mancha, Albacete, Spain;Computer Systems Department, Intelligent Systems & Data Mining--SIMD, I3A, University of Castilla-La Mancha, Albacete, Spain;Computer Systems Department, Intelligent Systems & Data Mining--SIMD, I3A, University of Castilla-La Mancha, Albacete, Spain

  • Venue:
  • Applied Intelligence
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Within the framework of Bayesian networks (BNs), most classifiers assume that the variables involved are of a discrete nature, but this assumption rarely holds in real problems. Despite the loss of information discretization entails, it is a direct easy-to-use mechanism that can offer some benefits: sometimes discretization improves the run time for certain algorithms; it provides a reduction in the value set and then a reduction in the noise which might be present in the data; in other cases, there are some Bayesian methods that can only deal with discrete variables. Hence, even though there are many ways to deal with continuous variables other than discretization, it is still commonly used. This paper presents a study of the impact of using different discretization strategies on a set of representative BN classifiers, with a significant sample consisting of 26 datasets. For this comparison, we have chosen Naive Bayes (NB) together with several other semi-Naive Bayes classifiers: Tree-Augmented Naive Bayes (TAN), k-Dependence Bayesian (KDB), Aggregating One-Dependence Estimators (AODE) and Hybrid AODE (HAODE). Also, we have included an augmented Bayesian network created by using a hill climbing algorithm (BNHC). With this comparison we analyse to what extent the type of discretization method affects classifier performance in terms of accuracy and bias-variance discretization. Our main conclusion is that even if a discretization method produces different results for a particular dataset, it does not really have an effect when classifiers are being compared. That is, given a set of datasets, accuracy values might vary but the classifier ranking is generally maintained. This is a very useful outcome, assuming that the type of discretization applied is not decisive future experiments can be d times faster, d being the number of discretization methods considered.