Learning Languages from Positive Data and Negative Counterexamples

In this paper we introduce a paradigm for learning in the limit of potentially infinite languages from all positive data and negative counterexamples provided in response to the conjectures made by the learner. Several variants of this paradigm are considered that reflect different conditions/constraints on the type and size of negative counterexamples and on the time for obtaining them. In particular, we consider the models where 1) a learner gets the least negative counterexample; 2) the size of a negative counterexample must be bounded by the size of the positive data seen so far; 3) a counterexample may be delayed. Learning power, limitations of these models, relationships between them, as well as their relationships with classical paradigms for learning languages in the limit (without negative counterexamples) are explored. Several surprising results are obtained. In particular, for Gold’s model of learning requiring a learner to syntactically stabilize on correct conjectures, learners getting negative counterexamples immediately turn out to be as powerful as the ones that do not get them for indefinitely (but finitely) long time (or are only told that their latest conjecture is not a subset of the target language, without any specific negative counterexample). Another result shows that for behaviourally correct learning (where semantic convergence is required from a learner) with negative counterexamples, a learner making just one error in almost all its conjectures has the “ultimate power”: it can learn the class of all recursively enumerable languages. Yet another result demonstrates that sometimes positive data and negative counterexamples provided by a teacher are not enough to compensate for full positive and negative data.

[1]  M. W. Shields An Introduction to Automata Theory , 1988 .

[2]  篠原 武 Studies on inductive inference from positive data , 1986 .

[3]  元木 達也 INDUCTIVE INFERENCE FROM ALL POSITIVE AND SOME NEGATIVE DATA , 1991 .

[4]  John Case,et al.  Machine Inductive Inference and Language Identification , 1982, ICALP.

[5]  William I. Gasarch,et al.  Learning via queries to an oracle , 1989, COLT '89.

[6]  Jeffrey D. Ullman,et al.  Introduction to Automata Theory, Languages and Computation , 1979 .

[7]  Manuel Blum,et al.  A Machine-Independent Theory of the Complexity of Recursive Functions , 1967, JACM.

[8]  Robert H. Sloan,et al.  BOOK REVIEW: "SYSTEMS THAT LEARN: AN INTRODUCTION TO LEARNING THEORY, SECOND EDITION", SANJAY JAIN, DANIEL OSHERSON, JAMES S. ROYER and ARUN SHARMA , 2001 .

[9]  Daniel N. Osherson,et al.  Systems That Learn: An Introduction to Learning Theory for Cognitive and Computer Scientists , 1990 .

[10]  R. Treiman,et al.  Brown & Hanlon revisited: mothers' sensitivity to ungrammatical forms , 1984, Journal of Child Language.

[11]  Thomas Zeugmann,et al.  Characterization of language learning front informant under various monotonicity constraints , 1994, J. Exp. Theor. Artif. Intell..

[12]  W. Gasarch,et al.  Bounded Queries in Recursion Theory , 1998 .

[13]  Jr. Hartley Rogers Theory of Recursive Functions and Effective Computability , 1969 .

[14]  John Case,et al.  Comparison of Identification Criteria for Machine Inductive Inference , 1983, Theor. Comput. Sci..

[15]  Stuart A. Kurtz,et al.  Extremes in the Degrees of Inferability , 1994, Ann. Pure Appl. Log..

[16]  Sandra Zilles,et al.  Learning languages with queries , 2002 .

[17]  C. Snow,et al.  Feedback to first language learners: the role of repetitions and clarification questions , 1986, Journal of Child Language.

[18]  Manuel Blum,et al.  Toward a Mathematical Theory of Inductive Inference , 1975, Inf. Control..

[19]  John Case,et al.  Language Learning with Some Negative Information , 1993, Journal of computer and system sciences (Print).

[20]  Dana Angluin,et al.  Queries and concept learning , 1988, Machine Learning.

[21]  E. Mark Gold,et al.  Language Identification in the Limit , 1967, Inf. Control..

[22]  Sandra Zilles,et al.  Comparison of Query Learning and Gold-Style Learning in Dependence of the Hypothesis Space , 2004, ALT.

[23]  Sandra Zilles,et al.  Replacing Limit Learners with Equally Powerful One-Shot Query Learners , 2004, COLT.

[24]  Paul Young,et al.  An introduction to the general theory of algorithms , 1978 .

[25]  Stephen A. Cook Review: Manuel Blum, A Machine-Independent Theory of the Complexity of Recursive Functions , 1969 .