How much complexity does an RNN architecture need to learn syntax-sensitive dependencies?

Long short-term memory (LSTM) networks and their variants are capable of encapsulating long-range dependencies, which is evident from their performance on a variety of linguistic tasks. On the other hand, simple recurrent networks (SRNs), which appear more biologically grounded in terms of synaptic connections, have generally been less successful at capturing long-range dependencies as well as the loci of grammatical errors in an unsupervised setting. In this paper, we seek to develop models that bridge the gap between biological plausibility and linguistic competence. We propose a new architecture, the Decay RNN, which incorporates the decaying nature of neuronal activations and models the excitatory and inhibitory connections in a population of neurons. Besides its biological inspiration, our model also shows competitive performance relative to LSTMs on subject-verb agreement, sentence grammaticality, and language modeling tasks. These results provide some pointers towards probing the nature of the inductive biases required for RNN architectures to model linguistic phenomena successfully.

[1]  Edouard Grave,et al.  Colorless Green Recurrent Networks Dream Hierarchically , 2018, NAACL.

[2]  Eleonora Catsigeras,et al.  Dale's Principle is necessary for an optimal neuronal network's dynamics , 2013, 1307.0597.

[3]  Surya Ganguli,et al.  On simplicity and complexity in the brave new world of large-scale neuroscience , 2015, Current Opinion in Neurobiology.

[4]  B Gluss,et al.  A model for neuron firing with exponential decay of potential resulting in diffusion equations for probability density. , 1967, The Bulletin of mathematical biophysics.

[5]  P. Strata,et al.  Dale’s principle , 1999, Brain Research Bulletin.

[6]  Hans J. Herrmann,et al.  Optimal percentage of inhibitory synapses in multi-task learning , 2015, Scientific Reports.

[7]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[8]  H. Mclennan,et al.  Synaptic Transmission , 2003 .

[9]  R. Thomas McCoy,et al.  Does Syntax Need to Grow on Trees? Sources of Hierarchical Inductive Bias in Sequence-to-Sequence Networks , 2020, TACL.

[10]  Niru Maheswaranathan,et al.  How recurrent networks implement contextual processing in sentiment analysis , 2020, ICML.

[11]  Jun Liang,et al.  Residual Recurrent Neural Networks for Learning Sequential Representations , 2018, Inf..

[12]  Christopher D. Manning,et al.  Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks , 2015, ACL.

[13]  Shuai Li,et al.  Independently Recurrent Neural Network (IndRNN): Building A Longer and Deeper RNN , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[14]  J. Eccles From electrical to chemical transmission in the central nervous system: The closing address of the Sir Henry Dale Centennial Symposium Cambridge, 19 September 1975 , 1976, Notes and Records of the Royal Society of London.

[15]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[16]  Xiao-Jing Wang,et al.  Task representations in neural networks trained to perform many cognitive tasks , 2019, Nature Neuroscience.

[17]  Guangyu R. Yang,et al.  Training Excitatory-Inhibitory Recurrent Neural Networks for Cognitive Tasks: A Simple and Flexible Framework , 2016, PLoS Comput. Biol..

[18]  Jürgen Schmidhuber,et al.  LSTM recurrent networks learn simple context-free and context-sensitive languages , 2001, IEEE Trans. Neural Networks.

[19]  Emmanuel Dupoux,et al.  Assessing the Ability of LSTMs to Learn Syntax-Sensitive Dependencies , 2016, TACL.

[20]  John Hale,et al.  LSTMs Can Learn Syntax-Sensitive Dependencies Well, But Modeling Structure Makes Them Better , 2018, ACL.

[21]  H. V. Gersdorff,et al.  Synaptic Transmission , 2008 .

[22]  E. Jones,et al.  Sizes and distributions of intrinsic neurons incorporating tritiated GABA in monkey sensory-motor cortex , 1981, The Journal of neuroscience : the official journal of the Society for Neuroscience.

[23]  H. Dale Pharmacology and Nerve-Endings , 1935 .

[24]  J. Knott The organization of behavior: A neuropsychological theory , 1951 .

[25]  Adam Santoro,et al.  Backpropagation and the brain , 2020, Nature Reviews Neuroscience.

[26]  Tal Linzen,et al.  Targeted Syntactic Evaluation of Language Models , 2018, EMNLP.

[27]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[28]  W. Newsome,et al.  Context-dependent computation by recurrent dynamics in prefrontal cortex , 2013, Nature.

[29]  Barnabás Póczos,et al.  The Statistical Recurrent Unit , 2017, ICML.

[30]  Razvan Pascanu,et al.  Advances in optimizing recurrent networks , 2012, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[31]  PHARMACOLOGY AND NERVE ENDINGS , 1934 .

[32]  Tommi S. Jaakkola,et al.  Tree-structured decoding with doubly-recurrent neural networks , 2016, ICLR.

[33]  Aaron C. Courville,et al.  Ordered Neurons: Integrating Tree Structures into Recurrent Neural Networks , 2018, ICLR.

[34]  Xiao-Jing Wang,et al.  Probabilistic Decision Making by Slow Reverberation in Cortical Circuits , 2002, Neuron.

[35]  Chris Dyer,et al.  A Critical Analysis of Biased Parsers in Unsupervised Parsing , 2019, ArXiv.