Iterative Learning from Texts and Counterexamples Using Additional Information
Document Type
Conference Proceeding
Publication Date
2009
Abstract
A variant of iterative learning in the limit (cf. [LZ96]) is studied when a learner gets negative examples refuting conjectures containing data in excess of the target language and uses additional information of the following four types: a) memorizing up to n input elements seen so far; b) up to n feedback memberships queries (testing if an item is a member of the input seen so far); c) the number of input elements seen so far; d) the maximal element of the input seen so far. We explore how additional information available to such learners (defined and studied in [JK07]) may help. In particular, we show that adding the maximal element or the number of elements seen so far helps such learners to infer any indexed class of languages class-preservingly (using a descriptive numbering defining the class) — as it is proved in [JK07], this is not possible without using additional information. We also study how, in the given context, different types of additional information fare against each other, and establish hierarchies of learners memorizing n + 1 versus n input elements seen and n + 1 versus n feedback membership queries.
DOI
10.1007/978-3-642-04414-4_26
Recommended Citation
Jain, S., & Kinber, E. (2009). Iterative learning from texts and counterexamples using additional information. In Gavaldà, R., Lugosi, G., Zeugmann, T., & Zilles, S. (Eds.), Algorithmic learning theory. ALT 2009. Springer. Doi: 10.1007/978-3-642-04414-4_26
Comments
ALT: International Conference on Algorithmic Learning Theory. 20th International Conference, Porto, Portugal, October 3-5, 2009.
Part of the series Lecture Notes in Computer Science, vol. 5809.