Comparing feedforward and recurrent neural network architectures with human behavior in artificial grammar learning

In recent years artificial neural networks achieved performance close to or better than humans in several domains: tasks that were previously human prerogatives, such as language processing, have witnessed remarkable improvements in state of the art models. One advantage of this technological boost is to facilitate comparison between different neural networks and human performance, in order to deepen our understanding of human cognition. Here, we investigate which neural network architecture (feedforward vs. recurrent) matches human behavior in artificial grammar learning, a crucial aspect of language acquisition. Prior experimental studies proved that artificial grammars can be learnt by human subjects after little exposure and often without explicit knowledge of the underlying rules. We tested four grammars with different complexity levels both in humans and in feedforward and recurrent networks. Our results show that both architectures can “learn” (via error back-propagation) the grammars after the same number of training sequences as humans do, but recurrent networks perform closer to humans than feedforward ones, irrespective of the grammar complexity level. Moreover, similar to visual processing, in which feedforward and recurrent architectures have been related to unconscious and conscious processes, the difference in performance between architectures over ten regular grammars shows that simpler and more explicit grammars are better learnt by recurrent architectures, supporting the hypothesis that explicit learning is best modeled by recurrent networks, whereas feedforward networks supposedly capture the dynamics involved in implicit learning.

[1]  Michael C. Mozer,et al.  A Focused Backpropagation Algorithm for Temporal Pattern Recognition , 1989, Complex Syst..

[2]  R. Sun,et al.  Learning in reactive sequential decision tasks: the CLARION model , 1996, Proceedings of International Conference on Neural Networks (ICNN'96).

[3]  Zoltan Dienes,et al.  Connectionist and Memory-Array Models of Artificial Grammar Learning , 1992, Cogn. Sci..

[4]  C. Koch,et al.  Neural correlates of consciousness: progress and problems , 2016, Nature Reviews Neuroscience.

[5]  Adrian M. Owen,et al.  The Quest for Consciousness , 2011 .

[6]  Axel Cleeremans,et al.  Unconscious associative learning with conscious cues , 2016, Neuroscience of consciousness.

[7]  David R. Shanks,et al.  Subjective measures of awareness and implicit cognition , 2003, Memory & cognition.

[8]  Jared G. Smith,et al.  The implicit sequence learning deficit in patients with Parkinson's disease: A matter of impaired sequence integration? , 2006, Neuropsychologia.

[9]  Y. Nesterov A method for solving the convex programming problem with convergence rate O(1/k^2) , 1983 .

[10]  Andrea Alamia,et al.  Which Neural Network Architecture matches Human Behavior in Artificial Grammar Learning? , 2019, 2019 Conference on Cognitive Computational Neuroscience.

[11]  Paul J. Werbos,et al.  Backpropagation Through Time: What It Does and How to Do It , 1990, Proc. IEEE.

[12]  A. Smith Grammar Inference Using Recurrent Neural Networks , 2022 .

[13]  D. Shanks IMPLICIT LEARNING AND TACIT KNOWLEDGE - AN ESSAY ON THE COGNITIVE UNCONSCIOUS - REBER,A , 1995 .

[14]  R. A. Carlson,et al.  A case of syntactical learning and judgment: How conscious and how abstract? , 1984 .

[15]  Aravind K. Joshi,et al.  Natural language parsing: Tree adjoining grammars: How much context-sensitivity is required to provide reasonable structural descriptions? , 1985 .

[16]  Karl J. Friston,et al.  Preserved Feedforward But Impaired Top-Down Processes in the Vegetative State , 2011, Science.

[17]  Roger Vergauwen Consciousness, Recursion and Language , 2014 .

[18]  Xi Zhang,et al.  The interaction of implicit learning, explicit hypothesis testing learning and implicit-to-explicit knowledge extraction , 2007, Neural Networks.

[19]  A. Turing On Computable Numbers, with an Application to the Entscheidungsproblem. , 1937 .

[20]  Mark Steedman,et al.  Plans, Affordances, And Combinatory Grammar , 2002 .

[21]  A. Buchner Indirect effects of synthetic grammar learning in an identification task. , 1994 .

[22]  Angela D. Friederici,et al.  Artificial grammar learning meets formal language theory: an overview , 2012, Philosophical Transactions of the Royal Society B: Biological Sciences.

[23]  Michael E J Masson,et al.  A tutorial on a practical Bayesian alternative to null-hypothesis significance testing , 2011, Behavior research methods.

[24]  James Rogers,et al.  Formal language theory: refining the Chomsky hierarchy , 2012, Philosophical Transactions of the Royal Society B: Biological Sciences.

[25]  Axel Cleeremans,et al.  Computational Models of Implicit Learning , 2019, Implicit Learning.

[26]  John R. Anderson,et al.  Learning Artificial Grammars With Competitive Chunking , 1990 .

[27]  Noam Chomsky,et al.  On Certain Formal Properties of Grammars , 1959, Inf. Control..

[28]  Zoltan Dienes,et al.  Two ways of learning associations , 2003, Cogn. Sci..

[29]  Axel Cleeremans,et al.  Implicit learning: news from the front , 1998, Trends in Cognitive Sciences.

[30]  Paul Rodríguez,et al.  A Recurrent Neural Network that Learns to Count , 1999, Connect. Sci..

[31]  Erik Cambria,et al.  Recent Trends in Deep Learning Based Natural Language Processing , 2017, IEEE Comput. Intell. Mag..

[32]  Peter A. Tass,et al.  Dendritic and Axonal Propagation Delays Determine Emergent Structures of Neuronal Networks with Plastic Synapses , 2017, Scientific Reports.

[33]  Ron Sun,et al.  Cognition and Multi-Agent Interaction: The CLARION Cognitive Architecture: Extending Cognitive Modeling to Social Simulation , 2005 .

[34]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[35]  Ron Sun,et al.  The Cambridge Handbook of Computational Psychology , 2008 .

[36]  V. Lamme,et al.  The distinct modes of vision offered by feedforward and recurrent processing , 2000, Trends in Neurosciences.

[37]  Jeffrey L. Elman,et al.  Finding Structure in Time , 1990, Cogn. Sci..

[38]  Annette Kinder,et al.  Connectionist models of artificial grammar learning: what type of knowledge is acquired? , 2009, Psychological research.

[39]  S. Thorpe,et al.  The Time Course of Visual Processing: From Early Perception to Decision-Making , 2001, Journal of Cognitive Neuroscience.

[40]  H. Spekreijse,et al.  Two distinct modes of sensory processing observed in monkey primary visual cortex (V1) , 2001, Nature Neuroscience.

[41]  Alaa A. Kharbouch,et al.  Three models for the description of language , 1956, IRE Trans. Inf. Theory.

[42]  A. Reber Implicit learning and tacit knowledge , 1993 .

[43]  Fang Zhao,et al.  Implicit sequence learning despite multitasking: the role of across-task predictability , 2019, Psychological research.

[44]  D. Shanks,et al.  Characteristics of dissociable human learning systems , 1994, Behavioral and Brain Sciences.

[45]  Larry R Squire,et al.  Conscious and Unconscious Memory Systems the Discovery of Declarative and Nondeclarative Memory Systems , 2022 .

[46]  D. Broadbent,et al.  On the Relationship between Task Performance and Associated Verbalizable Knowledge , 1984 .

[47]  James L. McClelland,et al.  Graded State Machines: The Representation of Temporal Contingencies in Simple Recurrent Networks , 2005, Machine Learning.

[48]  Martin Rohrmeier,et al.  Implicit Learning of Recursive Context-Free Grammars , 2012, PloS one.

[49]  Martin Rohrmeier,et al.  Implicit Learning and Recursion , 2014 .

[50]  Karl Magnus Petersson,et al.  Artificial grammar learning and neural networks , 2005 .

[51]  A. Reber,et al.  Another wrinkle on the dual-task SRT experiment: It’s probably not dual task , 2000, Psychonomic bulletin & review.

[52]  A. Reber Implicit learning of synthetic languages: The role of instructional set. , 1976 .

[53]  M. Abal,et al.  Determination of hemodynamic risk for vascular disease in planar artery bifurcations , 2018, Scientific Reports.

[54]  Rachel Schiff,et al.  Does complexity matter? Meta-analysis of learner performance in artificial grammar tasks , 2014, Front. Psychol..

[55]  E. Pothos Theories of artificial grammar learning. , 2007, Psychological bulletin.

[56]  Jonathan Berant,et al.  Inducing Regular Grammars Using Recurrent Neural Networks , 2017, ArXiv.

[57]  D. Hassabis,et al.  Neuroscience-Inspired Artificial Intelligence , 2017, Neuron.

[58]  Jürgen Schmidhuber,et al.  LSTM recurrent networks learn simple context-free and context-sensitive languages , 2001, IEEE Trans. Neural Networks.

[59]  Quan Wang,et al.  A model of human motor sequence learning explains facilitation and interference effects based on spike-timing dependent plasticity , 2017, PLoS Comput. Biol..

[60]  Paul Rodríguez,et al.  Simple Recurrent Networks Learn Context-Free and Context-Sensitive Languages by Counting , 2001, Neural Computation.

[61]  Z. Dienes,et al.  Implicit learning: Below the subjective threshold , 1997 .

[62]  Birgit Öttl,et al.  Does Formal Complexity Reflect Cognitive Complexity? Investigating Aspects of the Chomsky Hierarchy in an Artificial Language Learning Study , 2015, PloS one.

[63]  M. Wierzchoń,et al.  A Connectionist Model of Artificial Grammar Learning : Simulations Based on Higham ( 1997 ) Indexes of Knowledge Representation , 2008 .

[64]  C. Koch The Quest for Consciousness , 2004 .

[65]  Axel Cleeremans,et al.  Principles for Implicit Learning , 1997 .

[66]  Ray Jackendoff,et al.  Parallels and Nonparallels between Language and Music , 2009 .

[67]  Ray Jackendoff,et al.  What is the human language faculty?: Two views , 2011 .

[68]  R. VanRullen Perception Science in the Age of Deep Neural Networks , 2017, Front. Psychol..

[69]  A. Reber Implicit learning of artificial grammars , 1967 .

[70]  H. Schiöth,et al.  The role of G protein-coupled receptors in the early evolution of neurotransmission and the nervous system , 2015, Journal of Experimental Biology.

[71]  Axel Cleeremans Computational correlates of consciousness. , 2005, Progress in brain research.

[72]  L. Squire,et al.  Artificial grammar learning depends on implicit acquisition of both abstract and exemplar-specific information. , 1996, Journal of experimental psychology. Learning, memory, and cognition.

[73]  Z. Dienes,et al.  A theory of implicit and explicit knowledge , 1999, Behavioral and Brain Sciences.

[74]  C. Lee Giles,et al.  Learning and Extracting Finite State Automata with Second-Order Recurrent Neural Networks , 1992, Neural Computation.

[75]  S. Phillips,et al.  Processing capacity defined by relational complexity: implications for comparative, developmental, and cognitive psychology. , 1998, The Behavioral and brain sciences.

[76]  David R. Shanks,et al.  Amnesia and the Declarative/Nondeclarative Distinction: A Recurrent Network Model of Classification, Recognition, and Repetition Priming , 2001, Journal of Cognitive Neuroscience.

[77]  B. Newell,et al.  Degraded conditions: Confounds in the study of decision making , 2014, Behavioral and Brain Sciences.

[78]  John R. Vokey,et al.  Salience of Item Knowledge in Learning Artificial Grammars , 1992 .

[79]  W. Fitch,et al.  Self-similarity and recursion as default modes in human cognition , 2017, Cortex.

[80]  James L. McClelland,et al.  Learning the structure of event sequences. , 1991, Journal of experimental psychology. General.

[81]  Peter A. Tass,et al.  Delay-Induced Multistability and Loop Formation in Neuronal Networks with Spike-Timing-Dependent Plasticity , 2018, Scientific Reports.

[82]  Emil L. Post Recursively enumerable sets of positive integers and their decision problems , 1944 .

[83]  James L. McClelland,et al.  Finite State Automata and Simple Recurrent Networks , 1989, Neural Computation.

[84]  Jonathon Love,et al.  Software to Sharpen Your Stats , 2015 .

[85]  P. Frensch,et al.  Complex problem solving : the European perspective , 1995 .

[86]  Roger Payne,et al.  Large Scale Changes over 19 Years in Songs of Humpback Whales in Bermuda , 2010 .

[87]  Roger Vergauwen Consciousness, Recursion and Language (online december 2013) , 2014 .

[88]  Axel Cleeremans,et al.  Consciousness as a graded and an all-or-none phenomenon: A conceptual analysis , 2015, Consciousness and Cognition.

[89]  M. Corballis Recursive Cognition as a Prelude to Language , 2014 .

[90]  A. Reber,et al.  Implicit learning: An analysis of the form and structure of a body of tacit knowledge , 1977, Cognition.

[91]  Peggy Seriès,et al.  Self-Organized Artificial Grammar Learning in Spiking Neural Networks , 2014, CogSci.

[92]  John R. Vokey,et al.  Abstract analogies and abstracted grammars: Comments on Reber (1989) and Mathews et al. (1989). , 1991 .

[93]  Janet Wiles,et al.  Context-free and context-sensitive dynamics in recurrent neural networks , 2000, Connect. Sci..

[94]  J. DiCarlo,et al.  Using goal-driven deep learning models to understand sensory cortex , 2016, Nature Neuroscience.