Calibrated lazy associative classification

  • Authors:
  • Adriano Veloso;Wagner Meira, Jr.;Marcos Gonçalves;Humberto M. Almeida;Mohammed Zaki

  • Affiliations:
  • Computer Science Dept., Federal University of Minas Gerais, Brazil;Computer Science Dept., Federal University of Minas Gerais, Brazil;Computer Science Dept., Federal University of Minas Gerais, Brazil;Computer Science Dept., Federal University of Minas Gerais, Brazil;Computer Science Dept., Renseelaer Polytechnic Institute, USA

  • Venue:
  • Information Sciences: an International Journal
  • Year:
  • 2011

Quantified Score

Hi-index 0.07

Visualization

Abstract

Classification is a popular machine learning task. Given an example x and a class c, a classifier usually works by estimating the probability of x being member of c (i.e., membership probability). Well calibrated classifiers are those able to provide accurate estimates of class membership probabilities, that is, the estimated probability p@?(c|x) is close to p(c|p@?(c|x)), which is the true, (unknown) empirical probability of x being member of c given that the probability estimated by the classifier is p@?(c|x). Calibration is not a necessary property for producing accurate classifiers, and, thus, most of the research has focused on direct accuracy maximization strategies rather than on calibration. However, non-calibrated classifiers are problematic in applications where the reliability associated with a prediction must be taken into account. In these applications, a sensible use of the classifier must be based on the reliability of its predictions, and, thus, the classifier must be well calibrated. In this paper we show that lazy associative classifiers (LAC) are well calibrated using an MDL-based entropy minimization method. We investigate important applications where such characteristics (i.e., accuracy and calibration) are relevant, and we demonstrate empirically that LAC outperforms other classifiers, such as SVMs, Naive Bayes, and Decision Trees (even after these classifiers are calibrated). Additional highlights of LAC include the ability to incorporate reliable predictions for improving training, and the ability to refrain from doubtful predictions.