PAC learning with irrelevant attributes

  • Authors:
  • A. Dhagat;L. Hellerstein

  • Affiliations:
  • Dept. of Electr. Eng.&Comput. Sci., Wisconsin Univ., Milwaukee, WI, USA;-

  • Venue:
  • SFCS '94 Proceedings of the 35th Annual Symposium on Foundations of Computer Science
  • Year:
  • 1994

Quantified Score

Hi-index 0.00

Visualization

Abstract

We consider the problem of learning in the presence of irrelevant attributes in Valiant's PAC model (1984). In the PAC model, the goal of the learner is to produce an approximately correct hypothesis from random sample data. If the number of relevant attributes in the target function is small, it may be desirable to produce a hypothesis that also depends on only a small number of variables. Haussler (1988) previously considered the problem of learning monomials of a small number of variables. He showed that the greedy set cover approximation algorithm can be used as a polynomial-time Occam algorithm for learning monomials on r of n variables. A outputs a monomial on r(ln q+1) variables, where q is the number of negative examples in the sample. We extend this result by showing that there is a polynomial-time Occam algorithm for learning k-term DNF formulas depending on r of n variables that outputs a DNF formula depending on O(r/sup k/log/sup k/q) variables, where q is the number of negative examples in the sample. We also give a polynomial-time Occam algorithm for learning decision lists (sometimes called 1-decision lists) with k alternations.