Introducing Orthogonal Constraint in Structural Probes
暂无分享,去创建一个
[1] Alex Wang,et al. What do you learn from context? Probing for sentence structure in contextualized word representations , 2019, ICLR.
[2] Yonatan Belinkov,et al. Linguistic Knowledge and Transferability of Contextual Representations , 2019, NAACL.
[3] Anna Rumshisky,et al. A Primer in BERTology: What We Know About How BERT Works , 2020, Transactions of the Association for Computational Linguistics.
[4] Preslav Nakov,et al. Rotational Unit of Memory: A Novel Representation Unit for RNNs with Scalable Applications , 2019, TACL.
[5] Sampo Pyysalo,et al. Universal Dependencies v2: An Evergrowing Multilingual Treebank Collection , 2020, LREC.
[6] Emmanuel Dupoux,et al. Assessing the Ability of LSTMs to Learn Syntax-Sensitive Dependencies , 2016, TACL.
[7] Yonatan Belinkov,et al. What do Neural Machine Translation Models Learn about Morphology? , 2017, ACL.
[8] Zhangyang Wang,et al. Can We Gain More from Orthogonality Regularizations in Training Deep Networks? , 2018, NeurIPS.
[9] Les E. Atlas,et al. Full-Capacity Unitary Recurrent Neural Networks , 2016, NIPS.
[10] Dipanjan Das,et al. BERT Rediscovers the Classical NLP Pipeline , 2019, ACL.
[11] Quoc V. Le,et al. Exploiting Similarities among Languages for Machine Translation , 2013, ArXiv.
[12] Samuel R. Bowman,et al. A Gold Standard Dependency Corpus for English , 2014, LREC.
[13] E. L. Lehmann,et al. Introduction to Student (1908) The Probable Error of a Mean , 1992 .
[14] Anders Sogaard,et al. Are All Good Word Vector Spaces Isomorphic? , 2020, EMNLP.
[15] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[16] Omer Levy,et al. Deep RNNs Encode Soft Hierarchical Syntax , 2018, ACL.
[17] Eduard Hovy,et al. On the Systematicity of Probing Contextualized Word Representations: The Case of Hypernymy in BERT , 2020, STARSEM.
[18] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[19] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[20] Christopher D. Manning,et al. A Structural Probe for Finding Syntax in Word Representations , 2019, NAACL.
[21] Lysandre Debut,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[22] John Hewitt,et al. Designing and Interpreting Probes with Control Tasks , 2019, EMNLP.
[23] Pareto Probing: Trading Off Accuracy for Complexity , 2020, EMNLP.
[24] Christopher Joseph Pal,et al. On orthogonality and learning recurrent networks with long term dependencies , 2017, ICML.
[25] George A. Miller,et al. WordNet: A Lexical Database for English , 1995, HLT.
[26] Yoshua Bengio,et al. Unitary Evolution Recurrent Neural Networks , 2015, ICML.
[27] Yann LeCun,et al. Tunable Efficient Unitary Neural Networks (EUNN) and their application to RNNs , 2016, ICML.
[28] Ryan Cotterell,et al. Intrinsic Probing through Dimension Selection , 2020, EMNLP.
[29] Student,et al. THE PROBABLE ERROR OF A MEAN , 1908 .
[30] Samuel R. Bowman,et al. Language Modeling Teaches You More than Translation Does: Lessons Learned Through Auxiliary Syntactic Task Analysis , 2018, BlackboxNLP@EMNLP.
[31] Christopher D. Manning,et al. Finding Universal Grammatical Relations in Multilingual BERT , 2020, ACL.
[32] Joakim Nivre,et al. Do Neural Language Models Show Preferences for Syntactic Formalisms? , 2020, ACL.
[33] Yoshua Bengio,et al. Gated Orthogonal Recurrent Units: On Learning to Forget , 2017, Neural Computation.
[34] Yonatan Belinkov,et al. Analysis Methods in Neural Language Processing: A Survey , 2018, TACL.
[35] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.