Non-Disjoint discretization for aggregating one-dependence estimator classifiers

  • Authors:
  • Ana M. Martínez;Geoffrey I. Webb;M. Julia Flores;José A. Gámez

  • Affiliations:
  • Computer Systems Department, Intelligent Systems & Data Mining, University of Castilla-La Mancha, Albacete, Spain;Faculty of Information Technology, Monash University, Melbourne, Australia;Computer Systems Department, Intelligent Systems & Data Mining, University of Castilla-La Mancha, Albacete, Spain;Computer Systems Department, Intelligent Systems & Data Mining, University of Castilla-La Mancha, Albacete, Spain

  • Venue:
  • HAIS'12 Proceedings of the 7th international conference on Hybrid Artificial Intelligent Systems - Volume Part II
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

There is still lack of clarity about the best manner in which to handle numeric attributes when applying Bayesian network classifiers. Discretization methods entail an unavoidable loss of information. Nonetheless, a number of studies have shown that appropriate discretization can outperform straightforward use of common, but often unrealistic parametric distribution (e.g. Gaussian). Previous studies have shown the Averaged One-Dependence Estimators (AODE) classifier and its variant Hybrid AODE (HAODE, which deals with numeric and discrete variables) to be robust towards the discretization method applied. However, all the discretization techniques taken into account so far formed non-overlapping intervals for a numeric attribute. We argue that the idea of non-disjoint discretization, already justified in Naive Bayes classifiers, can also be profitably extended to AODE and HAODE, albeit with some variations; and our experimental results seem to support this hypothesis, specially for the latter.