Linguistic Knowledge and Transferability of Contextual Representations
暂无分享,去创建一个
Yonatan Belinkov | Noah A. Smith | Nelson F. Liu | Matt Gardner | Matthew E. Peters | Matt Gardner | Yonatan Belinkov
[1] Yoav Goldberg,et al. Coordination Annotation Extension in the Penn Tree Bank , 2016, ACL.
[2] Rico Sennrich,et al. Why Self-Attention? A Targeted Evaluation of Neural Machine Translation Architectures , 2018, EMNLP.
[3] Thorsten Brants,et al. One billion word benchmark for measuring progress in statistical language modeling , 2013, INTERSPEECH.
[4] Samuel R. Bowman,et al. Language Modeling Teaches You More than Translation Does: Lessons Learned Through Auxiliary Syntactic Task Analysis , 2018, BlackboxNLP@EMNLP.
[5] Holger Schwenk,et al. Supervised Learning of Universal Sentence Representations from Natural Language Inference Data , 2017, EMNLP.
[6] James Pustejovsky,et al. Are You Sure That This Happened? Assessing the Factuality Degree of Events in Text , 2012, CL.
[7] Fei-Fei Li,et al. Visualizing and Understanding Recurrent Networks , 2015, ArXiv.
[8] Richard Futrell,et al. Do RNNs learn human-like abstract word order preferences? , 2018, ArXiv.
[9] Alex Wang,et al. Looking for ELMo's friends: Sentence-Level Pretraining Beyond Language Modeling , 2018, ArXiv.
[10] Omer Levy,et al. Deep RNNs Encode Soft Hierarchical Syntax , 2018, ACL.
[11] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[12] Roger Levy,et al. What do RNN Language Models Learn about Filler–Gap Dependencies? , 2018, BlackboxNLP@EMNLP.
[13] Wang Ling,et al. Reference-Aware Language Models , 2016, EMNLP.
[14] Timothy Dozat,et al. Simpler but More Accurate Semantic Dependency Parsing , 2018, ACL.
[15] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[16] Sabine Buchholz,et al. Introduction to the CoNLL-2000 Shared Task Chunking , 2000, CoNLL/LLL.
[17] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[18] Stephan Oepen,et al. SemEval 2014 Task 8: Broad-Coverage Semantic Dependency Parsing , 2014, *SEMEVAL.
[19] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[20] Jungo Kasai,et al. Robust Multilingual Part-of-Speech Tagging via Adversarial Training , 2017, NAACL.
[21] Johan Bos,et al. The Parallel Meaning Bank: Towards a Multilingual Corpus of Translations Annotated with Compositional Meaning Representations , 2017, EACL.
[22] Noah A. Smith,et al. What Do Recurrent Neural Network Grammars Learn About Syntax? , 2016, EACL.
[23] Helen Yannakoudakis,et al. A New Dataset and Method for Automatically Grading ESOL Texts , 2011, ACL.
[24] James Pustejovsky,et al. FactBank: a corpus annotated with event factuality , 2009, Lang. Resour. Evaluation.
[25] Helen Yannakoudakis,et al. Compositional Sequence Labeling Models for Error Detection in Learner Writing , 2016, ACL.
[26] Yonatan Belinkov,et al. On internal language representations in deep learning: an analysis of machine translation and speech recognition , 2018 .
[27] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[28] Yuchen Zhang,et al. CoNLL-2012 Shared Task: Modeling Multilingual Unrestricted Coreference in OntoNotes , 2012, EMNLP-CoNLL Shared Task.
[29] Tal Linzen,et al. What can linguistics and deep learning contribute to each other? Response to Pater , 2018, Language.
[30] Omer Levy,et al. Simulating Action Dynamics with Neural Process Networks , 2017, ICLR.
[31] Guillaume Lample,et al. What you can cram into a single $&!#* vector: Probing sentence embeddings for linguistic properties , 2018, ACL.
[32] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[33] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[34] Sanja Fidler,et al. Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[35] Gonçalo Simões,et al. Morphosyntactic Tagging with a Meta-BiLSTM Model over Context Sensitive Token Encodings , 2018, ACL.
[36] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[37] Yejin Choi,et al. Dynamic Entity Representations in Neural Language Models , 2017, EMNLP.
[38] Dan Klein,et al. What’s Going On in Neural Constituency Parsers? An Analysis , 2018, NAACL.
[39] Regina Barzilay,et al. Cross-Lingual Alignment of Contextual Word Embeddings, with Applications to Zero-shot Dependency Parsing , 2019, NAACL.
[40] Rachel Rudinger,et al. Neural Models of Factuality , 2018, NAACL.
[41] Willem H. Zuidema,et al. Visualisation and 'diagnostic classifiers' reveal how recurrent and recursive neural networks process hierarchical structure , 2017, J. Artif. Intell. Res..
[42] Luke S. Zettlemoyer,et al. LSTM CCG Parsing , 2016, NAACL.
[43] Daniel Jurafsky,et al. Sharp Nearby, Fuzzy Far Away: How Neural Language Models Use Context , 2018, ACL.
[44] Mark Steedman,et al. CCGbank: A Corpus of CCG Derivations and Dependency Structures Extracted from the Penn Treebank , 2007, CL.
[45] Yonatan Belinkov,et al. Evaluating Layers of Representation in Neural Machine Translation on Part-of-Speech and Semantic Tagging Tasks , 2017, IJCNLP.
[46] Christopher Potts,et al. Did It Happen? The Pragmatic Complexity of Veridicality Assessment , 2012, CL.
[47] Srinivas Bangalore,et al. Supertagging: An Approach to Almost Parsing , 1999, CL.
[48] Emmanuel Dupoux,et al. Assessing the Ability of LSTMs to Learn Syntax-Sensitive Dependencies , 2016, TACL.
[49] Erik F. Tjong Kim Sang,et al. Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition , 2003, CoNLL.
[50] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[51] Dieuwke Hupkes,et al. Do Language Models Understand Anything? On the Ability of LSTMs to Understand Negative Polarity Items , 2018, BlackboxNLP@EMNLP.
[52] Alex Wang,et al. What do you learn from context? Probing for sentence structure in contextualized word representations , 2019, ICLR.
[53] Yonatan Belinkov,et al. What do Neural Machine Translation Models Learn about Morphology? , 2017, ACL.
[54] Nathan Schneider,et al. Comprehensive Supersense Disambiguation of English Prepositions and Possessives , 2018, ACL.
[55] Guillaume Lample,et al. Evaluation of Word Vector Representations by Subspace Alignment , 2015, EMNLP.
[56] Samuel R. Bowman,et al. A Gold Standard Dependency Corpus for English , 2014, LREC.
[57] Luke S. Zettlemoyer,et al. Dissecting Contextual Word Embeddings: Architecture and Representation , 2018, EMNLP.
[58] Ankur Bapna,et al. The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation , 2018, ACL.
[59] Noah A. Smith,et al. To Tune or Not to Tune? Adapting Pretrained Representations to Diverse Tasks , 2019, RepL4NLP@ACL.
[60] Yonatan Belinkov,et al. Fine-grained Analysis of Sentence Embeddings Using Auxiliary Prediction Tasks , 2016, ICLR.
[61] Luke S. Zettlemoyer,et al. AllenNLP: A Deep Semantic Natural Language Processing Platform , 2018, ArXiv.
[62] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[63] Anders Søgaard,et al. Jointly Learning to Label Sentences and Tokens , 2018, AAAI.
[64] Xinlei Chen,et al. Visualizing and Understanding Neural Models in NLP , 2015, NAACL.
[65] Yoshimasa Tsuruoka,et al. A Joint Many-Task Model: Growing a Neural Network for Multiple NLP Tasks , 2016, EMNLP.
[66] Timothy Dozat,et al. Deep Biaffine Attention for Neural Dependency Parsing , 2016, ICLR.
[67] Andrew McCallum,et al. Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.
[68] Yonatan Belinkov,et al. Analysis Methods in Neural Language Processing: A Survey , 2018, TACL.
[69] Xing Shi,et al. Does String-Based Neural MT Learn Source Syntax? , 2016, EMNLP.
[70] Alex Wang,et al. Can You Tell Me How to Get Past Sesame Street? Sentence-Level Pretraining Beyond Language Modeling , 2018, ACL.
[71] Deniz Yuret,et al. Why Neural Translations are the Right Length , 2016, EMNLP.
[72] Richard Socher,et al. Learned in Translation: Contextualized Word Vectors , 2017, NIPS.
[73] Johan Bos,et al. Semantic Tagging with Deep Residual Networks , 2016, COLING.