How much complexity does an RNN architecture need to learn syntax-sensitive dependencies?
暂无分享,去创建一个
Gantavya Bhatt | Hritik Bansal | Rishubh Singh | Sumeet Agarwal | Sumeet Agarwal | Hritik Bansal | Gantavya Bhatt | Rishu Singh
[1] Edouard Grave,et al. Colorless Green Recurrent Networks Dream Hierarchically , 2018, NAACL.
[2] Eleonora Catsigeras,et al. Dale's Principle is necessary for an optimal neuronal network's dynamics , 2013, 1307.0597.
[3] Surya Ganguli,et al. On simplicity and complexity in the brave new world of large-scale neuroscience , 2015, Current Opinion in Neurobiology.
[4] B Gluss,et al. A model for neuron firing with exponential decay of potential resulting in diffusion equations for probability density. , 1967, The Bulletin of mathematical biophysics.
[5] P. Strata,et al. Dale’s principle , 1999, Brain Research Bulletin.
[6] Hans J. Herrmann,et al. Optimal percentage of inhibitory synapses in multi-task learning , 2015, Scientific Reports.
[7] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[8] H. Mclennan,et al. Synaptic Transmission , 2003 .
[9] R. Thomas McCoy,et al. Does Syntax Need to Grow on Trees? Sources of Hierarchical Inductive Bias in Sequence-to-Sequence Networks , 2020, TACL.
[10] Niru Maheswaranathan,et al. How recurrent networks implement contextual processing in sentiment analysis , 2020, ICML.
[11] Jun Liang,et al. Residual Recurrent Neural Networks for Learning Sequential Representations , 2018, Inf..
[12] Christopher D. Manning,et al. Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks , 2015, ACL.
[13] Shuai Li,et al. Independently Recurrent Neural Network (IndRNN): Building A Longer and Deeper RNN , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[14] J. Eccles. From electrical to chemical transmission in the central nervous system: The closing address of the Sir Henry Dale Centennial Symposium Cambridge, 19 September 1975 , 1976, Notes and Records of the Royal Society of London.
[15] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[16] Xiao-Jing Wang,et al. Task representations in neural networks trained to perform many cognitive tasks , 2019, Nature Neuroscience.
[17] Guangyu R. Yang,et al. Training Excitatory-Inhibitory Recurrent Neural Networks for Cognitive Tasks: A Simple and Flexible Framework , 2016, PLoS Comput. Biol..
[18] Jürgen Schmidhuber,et al. LSTM recurrent networks learn simple context-free and context-sensitive languages , 2001, IEEE Trans. Neural Networks.
[19] Emmanuel Dupoux,et al. Assessing the Ability of LSTMs to Learn Syntax-Sensitive Dependencies , 2016, TACL.
[20] John Hale,et al. LSTMs Can Learn Syntax-Sensitive Dependencies Well, But Modeling Structure Makes Them Better , 2018, ACL.
[21] H. V. Gersdorff,et al. Synaptic Transmission , 2008 .
[22] E. Jones,et al. Sizes and distributions of intrinsic neurons incorporating tritiated GABA in monkey sensory-motor cortex , 1981, The Journal of neuroscience : the official journal of the Society for Neuroscience.
[23] H. Dale. Pharmacology and Nerve-Endings , 1935 .
[24] J. Knott. The organization of behavior: A neuropsychological theory , 1951 .
[25] Adam Santoro,et al. Backpropagation and the brain , 2020, Nature Reviews Neuroscience.
[26] Tal Linzen,et al. Targeted Syntactic Evaluation of Language Models , 2018, EMNLP.
[27] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[28] W. Newsome,et al. Context-dependent computation by recurrent dynamics in prefrontal cortex , 2013, Nature.
[29] Barnabás Póczos,et al. The Statistical Recurrent Unit , 2017, ICML.
[30] Razvan Pascanu,et al. Advances in optimizing recurrent networks , 2012, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[31] PHARMACOLOGY AND NERVE ENDINGS , 1934 .
[32] Tommi S. Jaakkola,et al. Tree-structured decoding with doubly-recurrent neural networks , 2016, ICLR.
[33] Aaron C. Courville,et al. Ordered Neurons: Integrating Tree Structures into Recurrent Neural Networks , 2018, ICLR.
[34] Xiao-Jing Wang,et al. Probabilistic Decision Making by Slow Reverberation in Cortical Circuits , 2002, Neuron.
[35] Chris Dyer,et al. A Critical Analysis of Biased Parsers in Unsupervised Parsing , 2019, ArXiv.