Iterative Learning from Positive Data and Negative Counterexamples

Document Type

Article

Publication Date

12-2007

Abstract

A model for learning in the limit is defined where a (so-called iterative) learner gets all positive examples from the target language, tests every new conjecture with a teacher (oracle) if it is a subset of the target language (and if it is not, then it receives a negative counterexample), and uses only limited long-term memory (incorporated in conjectures). Three variants of this model are compared: when a learner receives least negative counterexamples, the ones whose size is bounded by the maximum size of input seen so far, and arbitrary ones. A surprising result is that sometimes absence of bounded counterexamples can help an iterative learner whereas arbitrary counterexamples are useless. We also compare our learnability model with other relevant models of learnability in the limit, study how our model works for indexed classes of recursive languages, and show that learners in our model can work in non-U-shaped way—never abandoning the first right conjecture.

Comments

Published:

Jain, Sanjay and Efim Kinber. "Iterative Learning from Positive Data and Negative Counterexamples." Information and Computation 205.12 (2007): 1777–1805.

doi:10.1016/j.ic.2007.09.001


Share

COinS