Learning From Noisy Examples

  • Authors:
  • Dana Angluin;Philip Laird

  • Affiliations:
  • Department of Computer Science, Yale University, P.O. Box 2158, New Haven, CT 06520, U.S.A. ANGLUIN@YALE.EDU;NASA Ames Research Center, MS 244-17, Moffett Field, CA 94035, U.S.A. LAIRD%PLU@IO.ARC.NASA.GOV

  • Venue:
  • Machine Learning
  • Year:
  • 1988

Quantified Score

Hi-index 0.00

Visualization

Abstract

The basic question addressed in this paper is: how can a learning algorithm cope with incorrect training examples? Specifically, how can algorithms that produce an “approximately correct” identification with “high probability” for reliable data be adapted to handle noisy data? We show that when the teacher may make independent random errors in classifying the example data, the strategy of selecting the most consistent rule for the sample is sufficient, and usually requires a feasibly small number of examples, provided noise affects less than half the examples on average. In this setting we are able to estimate the rate of noise using only the knowledge that the rate is less than one half. The basic ideas extend to other types of random noise as well. We also show that the search problem associated with this strategy is intractable in general. However, for particular classes of rules the target rule may be efficiently identified if we use techniques specific to that class. For an important class of formulas – the k-CNF formulas studied by Valiant – we present a polynomial-time algorithm that identifies concepts in this form when the rate of classification errors is less than one half.