Learning Syntax by Automata Induction

In this paper we propose an explicit computer model for learning natural language syntax based on Angluin's (1982) efficient induction algorithms, using a complete corpus of grammatical example sentences. We use these results to show how inductive inference methods may be applied to learn substantial, coherent subparts of at least one natural language – English – that are not susceptible to the kinds of learning envisioned in linguistic theory. As two concrete case studies, we show how to learn English auxiliary verb sequences (such as could be taking, will have been taking) and the sequences of articles and adjectives that appear before noun phrases (such as the very old big deer). Both systems can be acquired in a computationally feasible amount of time using either positive examples, or, in an incremental mode, with implicit negative examples (examples outside a finite corpus are considered to be negative examples). As far as we know, this is the first computer procedure that learns a full-scale range of noun subclasses and noun phrase structure. The generalizations and the time required for acquisition match our knowledge of child language acquisition for these two cases. More importantly, these results show that just where linguistic theories admit to highly irregular subportions, we can apply efficient automata-theoretic learning algorithms. Since the algorithm works only for fragments of language syntax, we do not believe that it suffices for all of language acquisition. Rather, we would claim that language acquisition is nonuniform and susceptible to a variety of acquisition strategies; this algorithm may be one these.

[1]  Robert C. Berwick,et al.  Locality principles and the acquisition of syntactic knowledge , 1982 .

[2]  岩倉 国浩,et al.  2.英語の助動詞に関する1つのフィルター(第八部門「最近の助動詞研究-"The Category AUX in Universal Grammar"を中心に-」,日本英文学会第52回大会報告) , 1980 .

[3]  J. Gerard Wolff,et al.  Grammar Discovery as Data Compression , 1978, AISB/GI.

[4]  Robert C. Berwick,et al.  The acquisition of syntactic knowledge , 1985 .

[5]  Brian MacWhinney,et al.  Basic Syntactic Processes , 1982 .

[6]  Eric Wanner,et al.  Language acquisition: the state of the art , 1982 .

[7]  Tom Michael Mitchell Version spaces: an approach to concept learning. , 1979 .

[8]  R. Brown A First Language , 1973 .

[9]  Kenneth Wexler,et al.  Formal Principles of Language Acquisition , 1980 .

[10]  Dana Angluin,et al.  Inductive Inference of Formal Languages from Positive Data , 1980, Inf. Control..

[11]  Dana Angluin,et al.  Inference of Reversible Languages , 1982, JACM.

[12]  Steven Pinker,et al.  Language learnability and language development , 1985 .

[13]  J. Gerard Wolff,et al.  Language acquisition, data compression and generalization , 1982 .

[14]  Taylor L. Booth,et al.  Grammatical Inference: Introduction and Survey - Part I , 1975, IEEE Trans. Syst. Man Cybern..

[15]  Taylor L. Booth,et al.  Grammatical Inference: Introduction and Survey-Part I , 1986, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[16]  E. Mark Gold,et al.  Complexity of Automaton Identification from Given Data , 1978, Inf. Control..

[17]  Michael G. Thomason,et al.  Syntactic Pattern Recognition, An Introduction , 1978, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[18]  E. Mark Gold,et al.  Language Identification in the Limit , 1967, Inf. Control..