On the influence of imputation in classification: practical issues

  • Authors:
  • Eduardo R. Hruschka;Antonio J. T. Garcia;Estevam R. Hruschka, Jr;Nelson F. F. Ebecken

  • Affiliations:
  • Computer Science Department, University of Sao Paulo (USP), Sao Carlos, Brazil;IBM Software Group, Sao Paulo, Brazil;Computer Science Department, Federal University of Sao Carlos (UFSCAR), Sao Carlos, Brazil;COPPE, Federal University of Rio de Janeiro, Rio de Janeiro, Brazil

  • Venue:
  • Journal of Experimental & Theoretical Artificial Intelligence
  • Year:
  • 2009

Quantified Score

Hi-index 0.00

Visualization

Abstract

The substitution of missing values, also called imputation, is an important data preparation task for many domains. Ideally, the substitution of missing values should not insert biases into the dataset. This aspect has been usually assessed by some measures of the prediction capability of imputation methods. Such measures assume the simulation of missing entries for some attributes whose values are actually known. These artificially missing values are imputed and then compared with the original values. Although this evaluation is useful, it does not allow the influence of imputed values in the ultimate modelling task (e.g. in classification) to be inferred. We argue that imputation cannot be properly evaluated apart from the modelling task. Thus, alternative approaches are needed. This article elaborates on the influence of imputed values in classification. In particular, a practical procedure for estimating the inserted bias is described. As an additional contribution, we have used such a procedure to empirically illustrate the performance of three imputation methods (majority, naive Bayes and Bayesian networks) in three datasets. Three classifiers (decision tree, naive Bayes and nearest neighbours) have been used as modelling tools in our experiments. The achieved results illustrate a variety of situations that can take place in the data preparation practice.