Cognition and the computational power of connectionist networks

This paper examines certain claims of 'cognitive significance' which (wisely or not) have been based upon the theoretical powers of three distinct classes of connectionist networks, namely, the 'universal function approximators', recurrent finite-state simulation networks and Turing equivalent networks. Each class will be considered with respect to its potential in the realm of cognitive modelling. Regarding the first class, I argue that, contrary to the claims of some influential connectionists, feed-forward networks do not possess the theoretical capacity to approximate all functions of interest to cognitive scientists. For example, they cannot approximate many important, recursive (halting) functions which map symbolic strings onto other symbolic strings. By contrast, I argue that a certain class of recurrent networks (i.e. those which closely approximate deterministic finite automata (DFA)) shows considerably greater promise in some domains. However, from a cognitive standpoint, difficulties arise when we consider how the relevant recurrent networks could acquire the weight vectors needed to support DFA simulations. These difficulties are severe in the realm of central high-level cognitive functions. In addition, the class of Turing equivalent networks is here examined. It is argued that the relevance ofsuch networks to cognitive modelling is seriously undermined by their reliance on infinite precision in crucial weights and/or node activations. I also examine what advantages these networks might conceivably possess over and above classical symbolic algorithms. For, from a cognitive standpoint, the Turing equivalent networks present difficulties very similar to certain classical algorithms; they appear highly contrived, their structure is fragile and they exhibit little or no noise tolerance.

[1]  Hava T. Siegelmann,et al.  The Dynamic Universality of Sigmoidal Neural Networks , 1996, Inf. Comput..

[2]  James L. McClelland,et al.  Finite State Automata and Simple Recurrent Networks , 1989, Neural Computation.

[3]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[4]  G. Marcus Rethinking Eliminative Connectionism , 1998, Cognitive Psychology.

[5]  T. Gelder,et al.  On Being Systematically Connectionist , 1994 .

[6]  Mike Casey,et al.  The Dynamics of Discrete-Time Computation, with Application to Recurrent Neural Networks and Finite State Machine Extraction , 1996, Neural Computation.

[7]  L. Shastri,et al.  From simple associations to systematic reasoning: A connectionist representation of rules, variables and dynamic bindings using temporal synchrony , 1993, Behavioral and Brain Sciences.

[8]  Jeffrey L. Elman,et al.  Finding Structure in Time , 1990, Cogn. Sci..

[9]  Hava T. Siegelmann,et al.  On the Computational Power of Neural Nets , 1995, J. Comput. Syst. Sci..

[10]  P. Thagard,et al.  Explanatory coherence , 1993 .

[11]  Hava T. Siegelmann,et al.  The Simple Dynamics of Super Turing Theories , 1996, Theor. Comput. Sci..

[12]  Robert F. Hadley,et al.  Language acquisition from sparse input without error feedback , 1999, Neural Networks.

[13]  Albert S. Bregman,et al.  Imagery and language acquisition , 1973 .

[14]  James D. Keeler,et al.  Layered Neural Networks with Gaussian Hidden Units as Universal Approximations , 1990, Neural Computation.

[15]  Allen Newell,et al.  SOAR: An Architecture for General Intelligence , 1987, Artif. Intell..

[16]  Charles X. Ling,et al.  Learning the Past Tense of English Verbs: The Symbolic Pattern Associator vs. Connectionist Models , 1993, J. Artif. Intell. Res..

[17]  J. Fodor,et al.  Connectionism and cognitive architecture: A critical analysis , 1988, Cognition.

[18]  J. Elman Learning and development in neural networks: the importance of starting small , 1993, Cognition.