暂无分享,去创建一个
[1] Samuel R. Bowman,et al. CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models , 2020, EMNLP.
[2] Chandler May,et al. On Measuring Social Biases in Sentence Encoders , 2019, NAACL.
[3] M. Banaji,et al. Implicit social cognition: attitudes, self-esteem, and stereotypes. , 1995, Psychological review.
[4] Rachel Rudinger,et al. Gender Bias in Coreference Resolution , 2018, NAACL.
[5] Emily M. Bender,et al. On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜 , 2021, FAccT.
[6] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[7] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[8] Saif Mohammad,et al. Examining Gender and Race Bias in Two Hundred Sentiment Analysis Systems , 2018, *SEMEVAL.
[9] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[10] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[11] Alan W Black,et al. Measuring Bias in Contextualized Word Representations , 2019, Proceedings of the First Workshop on Gender Bias in Natural Language Processing.
[12] A. Tversky,et al. Judgment under Uncertainty: Heuristics and Biases , 1974, Science.
[13] Adam Kilgarriff,et al. The TenTen Corpus Family , 2013 .
[14] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[15] Alan W Black,et al. Black is to Criminal as Caucasian is to Police: Detecting and Removing Multiclass Bias in Word Embeddings , 2019, NAACL.
[16] Davis Liang,et al. Masked Language Model Scoring , 2019, ACL.
[17] Arvind Narayanan,et al. Semantics derived automatically from language corpora contain human-like biases , 2016, Science.
[18] Brian A. Nosek,et al. Math = male, me = female, therefore math ≠ me. , 2002 .
[19] Panagiotis G. Ipeirotis,et al. Demographics and Dynamics of Mechanical Turk Workers , 2018, WSDM.
[20] Yi Tay,et al. How Reliable are Model Diagnostics? , 2021, FINDINGS.
[21] Jieyu Zhao,et al. Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods , 2018, NAACL.
[22] Markus Krötzsch,et al. Wikidata , 2014, Commun. ACM.
[23] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[24] A. Kilgarriff. Simple Maths for Keywords , 2009 .
[25] Nanyun Peng,et al. The Woman Worked as a Babysitter: On Biases in Language Generation , 2019, EMNLP.
[26] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[27] Sapna Cheryan,et al. Positive Stereotypes Are Pervasive and Powerful , 2015, Perspectives on psychological science : a journal of the Association for Psychological Science.
[28] G. N. Rider,et al. Black sexual politics: African Americans, gender, and the new racism , 2014, Culture, health & sexuality.
[29] Brian A. Nosek,et al. Math Male , Me Female , Therefore Math Me , 2002 .
[30] Yoav Goldberg,et al. Lipstick on a Pig: Debiasing Methods Cover up Systematic Gender Biases in Word Embeddings But do not Remove Them , 2019, NAACL-HLT.
[31] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[32] Sanja Fidler,et al. Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[33] Adam Tauman Kalai,et al. Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings , 2016, NIPS.
[34] Emily M. Bender,et al. Data Statements for Natural Language Processing: Toward Mitigating System Bias and Enabling Better Science , 2018, TACL.
[35] Christopher Potts,et al. Learning Word Vectors for Sentiment Analysis , 2011, ACL.