Optimal Language Learning

Gold's original paper on inductive inference introduced a notion of an optimal learner. Intuitively, a learner identifies a class of objects optimally iff there is no otherlearner that: requires as littleof each presentation of each object in the class in order to identify that object, and, for somepresentation of someobject in the class, requires lessof that presentation in order to identify that object. Wiehagen considered this notion in the context of functionlearning, and characterizedan optimal function learner as one that is class-preserving, consistent, and (in a very strong sense) non-U-shaped, with respect to the class of functions learned. Herein, Gold's notion is considered in the context of languagelearning. Intuitively, a language learner identifies a class of languages optimally iff there is no other learner that: requires as little of each textfor each language in the class in order to identify that language, and, for some text for some language in the class, requires less of that text in order to identify that language. Many interesting results concerning optimal language learners are presented. First, it is shown that a characterization analogous to Wiehagen's does nothold in this setting. Specifically, optimality is notsufficient to guarantee Wiehagen's conditions; though, those conditions aresufficient to guarantee optimality. Second, it is shown that the failure of this analog is notdue to a restriction on algorithmic learning power imposed by non-U-shapedness (in the strong form employed by Wiehagen). That is, non-U-shapedness, even in this strong form, does notrestrict algorithmic learning power. Finally, for an arbitrary optimal learner F of a class of languages $\mathcal {L}$, it is shown that F optimally identifies a subclass $\mathcal {K}$ of $\mathcal {L}$ iff F is class-preserving with respect to $\mathcal {K}$.

[1]  John Case,et al.  Comparison of Identification Criteria for Machine Inductive Inference , 1983, Theor. Comput. Sci..

[2]  Jr. Hartley Rogers Theory of Recursive Functions and Effective Computability , 1969 .

[3]  John Case,et al.  U-shaped, iterative, and iterative-with-counter learning , 2007, Machine Learning.

[4]  Daniel N. Osherson,et al.  Systems That Learn: An Introduction to Learning Theory for Cognitive and Computer Scientists , 1990 .

[5]  John Case,et al.  Results on memory-limited U-shaped learning , 2007, Inf. Comput..

[6]  Rolf Wiehagen A Thesis in Inductive Inference , 1990, Nonmonotonic and Inductive Logic.

[7]  E. Mark Gold,et al.  Language Identification in the Limit , 1967, Inf. Control..

[8]  Dana Angluin,et al.  Inductive Inference of Formal Languages from Positive Data , 1980, Inf. Control..

[9]  Mark A. Fulk Prudence and Other Conditions on Formal Language Learning , 1990, Inf. Comput..

[10]  Lorenzo Carlucci,et al.  Non-U-shaped vacillatory and team learning , 2008, J. Comput. Syst. Sci..

[11]  Robert H. Sloan,et al.  BOOK REVIEW: "SYSTEMS THAT LEARN: AN INTRODUCTION TO LEARNING THEORY, SECOND EDITION", SANJAY JAIN, DANIEL OSHERSON, JAMES S. ROYER and ARUN SHARMA , 2001 .

[12]  John Case,et al.  When unlearning helps , 2008, Inf. Comput..

[13]  Leonard Pitt,et al.  Inductive Inference, DFAs, and Computational Complexity , 1989, AII.

[14]  Manuel Blum,et al.  Toward a Mathematical Theory of Inductive Inference , 1975, Inf. Control..