Iterative learning from positive data and negative counterexamples

  • Authors:
  • Sanjay Jain;Efim Kinber

  • Affiliations:
  • School of Computing, National University of Singapore, Singapore;Department of Computer Science, Sacred Heart University, Fairfield, CT

  • Venue:
  • ALT'06 Proceedings of the 17th international conference on Algorithmic Learning Theory
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

A model for learning in the limit is defined where a (so-called iterative) learner gets all positive examples from the target language, tests every new conjecture with a teacher (oracle) if it is a subset of the target language (and if it is not, then it receives a negative counterexample), and uses only limited long-term memory (incorporated in conjectures). Three variants of this model are compared: when a learner receives least negative counterexamples, the ones whose size is bounded by the maximum size of input seen so far, and arbitrary ones. We also compare our learnability model with other relevant models of learnability in the limit, study how our model works for indexed classes of recursive languages, and show that learners in our model can work in non-U-shaped way — never abandoning the first right conjecture.