How to Design a Connectionist Holistic Parser

Connectionist holistic parsing offers a viable and attractive alternative to traditional algorithmic parsers. With exposure to a limited subset of grammatical sentences and their corresponding parse trees only, a holistic parser is capable of learning inductively the grammatical regularity underlying the training examples that affects the parsing process. In the past, various connectionist parsers have been proposed. Each approach had its own unique characteristics, and yet some techniques were shared in common. In this article, various dimensions underlying the design of a holistic parser are explored, including the methods to encode sentences and parse trees, whether a sentence and its corresponding parse tree share the same representation, the use of confluent inference, and the inclusion of phrases in the training set. Different combinations of these design factors give rise to different holistic parsers. In succeeding discussions, we scrutinize these design techniques and compare the performances of a few parsers on language parsing, including the confluent preorder parser, the backpropagation parsing network, the XERIC parser of Berg (1992), the modular connectionist parser of Sharkey and Sharkey (1992), Reilly's (1992) model, and their derivatives. Experiments are performed to evaluate their generalization capability and robustness. The results reveal a number of issues essential for building an effective holistic parser.

[1]  Mitchell P. Marcus,et al.  A theory of syntactic recognition for natural language , 1979 .

[2]  J. Fodor,et al.  Connectionism and cognitive architecture: A critical analysis , 1988, Cognition.

[3]  Peter Tiño,et al.  Learning long-term dependencies in NARX recurrent neural networks , 1996, IEEE Trans. Neural Networks.

[4]  Douglas S. Blank,et al.  Exploring the Symbolic/Subsymbolic Continuum: A case study of RAAM , 1992 .

[5]  Dominic Palmer-Brown,et al.  (S)RAAM: An Analytical Technique for Fast and Reliable Derivation of Connectionist Symbol Structure Representations , 1997, Connect. Sci..

[6]  Barry L. Kalman,et al.  Tail-recursive Distributed Representations and Simple Recurrent Networks , 1995 .

[7]  James Alistair Hammerton,et al.  Holistic Computation: Reconstructing a Muddled Concept , 1998, Connect. Sci..

[8]  Noel E. Sharkey,et al.  Connectionist Natural Language Processing , 1992 .

[9]  B. Apolloni,et al.  A general framework for symbol and rule extraction in neural networks , 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on Neural Networks. IJCNN 2000. Neural Computing: New Challenges and Perspectives for the New Millennium.

[10]  Alfred V. Aho,et al.  Compilers: Principles, Techniques, and Tools , 1986, Addison-Wesley series in computer science / World student series edition.

[11]  Andreas Stolcke,et al.  Tree matching with recursive distributed representations , 1992, AAAI Conference on Artificial Intelligence.

[12]  Ronan G. Reilly,et al.  Connectionist technique for on-line parsing , 1992 .

[13]  Alexander Franz Learning PP attachment from corpus statistics , 1995, Learning for Natural Language Processing.

[14]  George Berg,et al.  A Connectionist Parser with Recursive Sentence Structure and Lexical Disambiguation , 1992, AAAI.

[15]  Jordan B. Pollack,et al.  Massively Parallel Parsing: A Strongly Interactive Model of Natural Language Interpretation , 1988, Cogn. Sci..

[16]  James F. Allen Natural language understanding , 1987, Bejnamin/Cummings series in computer science.

[17]  Jeffrey L. Elman,et al.  Finding Structure in Time , 1990, Cogn. Sci..

[18]  David W. Carroll,et al.  Psychology of Language , 1993 .

[19]  Chris Mellish,et al.  Natural Language Processing in Pop-11: An Introduction to Computational Linguistics , 1989 .

[20]  Gerald Gazdar,et al.  Natural Language Processing in PROLOG: An Introduction to Computational Linguistics , 1989 .

[21]  Gilbert K. Krulee,et al.  Computer processing of natural language , 1985 .

[22]  Eugene Charniak,et al.  Statistical language learning , 1997 .

[23]  Lonnie Chrisman,et al.  Learning Recursive Distributed Representations for Holistic Computation , 1991 .

[24]  Andrew S. Noetzel,et al.  Forced Simple Recurrent Neural Networks and Grammatical Inference , 1992 .

[25]  James F. Allen Natural language understanding (2nd ed.) , 1995 .

[26]  Ho Kei Shin Edward,et al.  Representing Sentence Structures on Neural Networks , 1994 .

[27]  Risto Miikkulainen Subsymbolic Case-Role Analysis of Sentences with Embedded Clauses , 1993 .

[28]  Ajay N. Jain Parsing Complex Sentences with Structured Connectionist Networks , 1991, Neural Computation.

[29]  Jordan B. Pollack,et al.  Recursive Distributed Representations , 1990, Artif. Intell..

[30]  Kanaan A. Faisal,et al.  Symbolic parsing via subsymbolic rules , 1992 .

[31]  Lai-Wan Chan,et al.  Confluent Preorder Parsing of Deterministic Grammars , 1997, Connect. Sci..