暂无分享,去创建一个
Leon Bergen | Charles Yu | Ryan Sie | Nico Tedeschi | Leon Bergen | N. Tedeschi | Charles Yu | Ryan Sie
[1] Michael A. Lepori,et al. Representations of Syntax [MASK] Useful: Effects of Constituency and Dependency Structure in Recursive LSTMs , 2020, ACL.
[2] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[3] Edouard Grave,et al. Colorless Green Recurrent Networks Dream Hierarchically , 2018, NAACL.
[4] Florian Mohnert,et al. Under the Hood: Using Diagnostic Classifiers to Investigate and Improve how Language Models Track Agreement Information , 2018, BlackboxNLP@EMNLP.
[5] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[6] Fred Landman. Distributivity, Collectivity and Cumulativity , 2000 .
[7] Tal Linzen,et al. Targeted Syntactic Evaluation of Language Models , 2018, EMNLP.
[8] Roger Levy,et al. What do RNN Language Models Learn about Filler–Gap Dependencies? , 2018, BlackboxNLP@EMNLP.
[9] Yoav Goldberg,et al. Assessing BERT's Syntactic Abilities , 2019, ArXiv.
[10] Yiming Yang,et al. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context , 2019, ACL.
[11] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[12] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[13] John Hale,et al. LSTMs Can Learn Syntax-Sensitive Dependencies Well, But Modeling Structure Makes Them Better , 2018, ACL.
[14] Peter Lasersohn. Mass Nouns and Plurals , 2011 .
[15] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[16] R Ratcliff,et al. Connectionist models of recognition memory: constraints imposed by learning and forgetting functions. , 1990, Psychological review.
[17] R. French. Catastrophic forgetting in connectionist networks , 1999, Trends in Cognitive Sciences.
[18] Christiane Fellbaum,et al. Book Reviews: WordNet: An Electronic Lexical Database , 1999, CL.
[19] Roger Levy,et al. Structural Supervision Improves Learning of Non-Local Grammatical Dependencies , 2019, NAACL.
[20] Jan Tore Lønning,et al. Plurals and Collectivity , 1997, Handbook of Logic and Language.
[21] Emmanuel Dupoux,et al. Assessing the Ability of LSTMs to Learn Syntax-Sensitive Dependencies , 2016, TACL.
[22] Alexander Clark,et al. Grammaticality, Acceptability, and Probability: A Probabilistic View of Linguistic Knowledge , 2017, Cogn. Sci..
[23] Beth Levin,et al. English Verb Classes and Alternations: A Preliminary Investigation , 1993 .
[24] Samuel R. Bowman,et al. Neural Network Acceptability Judgments , 2018, Transactions of the Association for Computational Linguistics.
[25] Shikha Bordia,et al. Investigating BERT’s Knowledge of Language: Five Analysis Methods with NPIs , 2019, EMNLP.
[26] R'emi Louf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[27] Roger Levy,et al. Neural language models as psycholinguistic subjects: Representations of syntactic state , 2019, NAACL.
[28] Martha Palmer,et al. Verbnet: a broad-coverage, comprehensive verb lexicon , 2005 .