Controlling the Imprint of Passivization and Negation in Contextualized Representations
暂无分享,去创建一个
Jörg Tiedemann | Marianna Apidianaki | Hande Celikkanat | Sami Virpioja | H. Çelikkanat | J. Tiedemann | Sami Virpioja | Marianna Apidianaki
[1] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[2] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[3] Fedor Moiseev,et al. Analyzing Multi-Head Self-Attention: Specialized Heads Do the Heavy Lifting, the Rest Can Be Pruned , 2019, ACL.
[4] Yoav Goldberg,et al. Adversarial Removal of Demographic Attributes from Text Data , 2018, EMNLP.
[5] Tal Linzen,et al. Using Priming to Uncover the Organization of Syntactic Representations in Neural Language Models , 2019, CoNLL.
[6] Anna Rumshisky,et al. A Primer in BERTology: What We Know About How BERT Works , 2020, Transactions of the Association for Computational Linguistics.
[7] Yoav Goldberg,et al. Null It Out: Guarding Protected Attributes by Iterative Nullspace Projection , 2020, ACL.
[8] Dipanjan Das,et al. BERT Rediscovers the Classical NLP Pipeline , 2019, ACL.
[9] Omer Levy,et al. What Does BERT Look at? An Analysis of BERT’s Attention , 2019, BlackboxNLP@ACL.
[10] Rico Sennrich,et al. The Bottom-up Evolution of Representations in the Transformer: A Study with Machine Translation and Language Modeling Objectives , 2019, EMNLP.
[11] Goran Glavas,et al. A General Framework for Implicit and Explicit Debiasing of Distributional Word Vector Spaces , 2020, AAAI.
[12] Marco Marelli,et al. A SICK cure for the evaluation of compositional distributional semantic models , 2014, LREC.
[13] M. Marelli,et al. SemEval-2014 Task 1: Evaluation of Compositional Distributional Semantic Models on Full Sentences through Semantic Relatedness and Textual Entailment , 2014, *SEMEVAL.
[14] Christopher D. Manning,et al. Stanza: A Python Natural Language Processing Toolkit for Many Human Languages , 2020, ACL.
[15] Adam Tauman Kalai,et al. Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings , 2016, NIPS.
[16] Holger Schwenk,et al. Massively Multilingual Sentence Embeddings for Zero-Shot Cross-Lingual Transfer and Beyond , 2018, Transactions of the Association for Computational Linguistics.
[17] Christopher D. Manning,et al. A Structural Probe for Finding Syntax in Word Representations , 2019, NAACL.
[18] Jörg Tiedemann,et al. A Systematic Study of Inner-Attention-Based Sentence Representations in Multilingual Neural Machine Translation , 2020, CL.
[19] Jonathan Berant,et al. oLMpics-On What Language Model Pre-training Captures , 2019, Transactions of the Association for Computational Linguistics.
[20] Ke Xu,et al. Cleaning the Null Space: A Privacy Mechanism for Predictors , 2017, AAAI.
[21] Philip Resnik,et al. More than Words: Syntactic Packaging and Implicit Sentiment , 2009, NAACL.
[22] Graham Neubig,et al. Controllable Invariance through Adversarial Feature Learning , 2017, NIPS.
[23] Emmanuel Dupoux,et al. Assessing the Ability of LSTMs to Learn Syntax-Sensitive Dependencies , 2016, TACL.
[24] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[25] John Hewitt,et al. Designing and Interpreting Probes with Control Tasks , 2019, EMNLP.
[26] Jeff M. Phillips,et al. Attenuating Bias in Word Vectors , 2019, AISTATS.
[27] Jörg Tiedemann,et al. Are Multilingual Neural Machine Translation Models Better at Capturing Linguistic Features? , 2020, Prague Bull. Math. Linguistics.
[28] Anna Rumshisky,et al. Revealing the Dark Secrets of BERT , 2019, EMNLP.
[29] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[30] Zeyu Li,et al. Learning Gender-Neutral Word Embeddings , 2018, EMNLP.
[31] Philipp Koehn,et al. Europarl: A Parallel Corpus for Statistical Machine Translation , 2005, MTSUMMIT.
[32] Martin Wattenberg,et al. Google’s Multilingual Neural Machine Translation System: Enabling Zero-Shot Translation , 2016, TACL.
[33] Blake Lemoine,et al. Mitigating Unwanted Biases with Adversarial Learning , 2018, AIES.