Having Beer after Prayer? Measuring Cultural Bias in Large Language Models
暂无分享,去创建一个
[1] Daniel Hershcovich,et al. Assessing Cross-Cultural Alignment between ChatGPT and Human Societies: An Empirical Study , 2023, C3NLP.
[2] Parishad BehnamGhader,et al. An Analysis of Social Biases Present in BERT Variants Across Multiple Languages , 2022, ArXiv.
[3] Alexander M. Rush,et al. BLOOM: A 176B-Parameter Open-Access Multilingual Language Model , 2022, ArXiv.
[4] Dan Jurafsky,et al. Multilingual BERT has an accent: Evaluating English influences on fluency in multilingual models , 2022, FINDINGS.
[5] Partha P. Talukdar,et al. Re-contextualizing Fairness in NLP: The Case of India , 2022, AACL.
[6] Yang Trista Cao,et al. Theory-Grounded Measurement of U.S. Social Stereotypes in English Language Models , 2022, NAACL.
[7] Kathleen C. Fraser,et al. Does Moral Code have a Moral Code? Probing Delphi’s Moral Philosophy , 2022, TRUSTNLP.
[8] Eric Michael Smith,et al. “I’m sorry to hear that”: Finding New Biases in Language Models with a Holistic Descriptor Dataset , 2022, EMNLP.
[9] Naoaki Okazaki,et al. Gender Bias in Masked Language Models for Multiple Languages , 2022, NAACL.
[10] Tatiana Shavrina,et al. mGPT: Few-Shot Learners Go Multilingual , 2022, ArXiv.
[11] Yadollah Yaghoobzadeh,et al. Metaphors in Pre-Trained Language Models: Probing and Generalization Across Datasets and Languages , 2022, ACL.
[12] Noah A. Smith,et al. Annotators with Attitudes: How Annotator Beliefs And Identities Bias Toxic Language Detection , 2021, NAACL.
[13] Alice H. Oh,et al. Mitigating Language-Dependent Ethnic Bias in BERT , 2021, EMNLP.
[14] Kashif Shah,et al. Quantifying Social Biases in NLP: A Generalization and Empirical Comparison of Extrinsic Fairness Metrics , 2021, Transactions of the Association for Computational Linguistics.
[15] Dirk Hovy,et al. HONEST: Measuring Hurtful Sentence Completion in Language Models , 2021, NAACL.
[16] Abubakar Abid,et al. Large language models associate Muslims with violence , 2021, Nature Machine Intelligence.
[17] C. Rothkopf,et al. Large pre-trained language models contain human-like biases of what is right and wrong to do , 2021, Nature Machine Intelligence.
[18] Arkaitz Zubiaga,et al. Towards generalisable hate speech detection: a review on obstacles and solutions , 2021, PeerJ Comput. Sci..
[19] Yejin Choi,et al. Challenges in Automated Debiasing for Toxic Language Detection , 2021, EACL.
[20] James Zou,et al. Persistent Anti-Muslim Bias in Large Language Models , 2021, AIES.
[21] Hazem M. Hajj,et al. AraGPT2: Pre-Trained Transformer for Arabic Language Generation , 2020, WANLP.
[22] Iryna Gurevych,et al. UNKs Everywhere: Adapting Multilingual Language Models to New Scripts , 2020, EMNLP.
[23] Adam Lopez,et al. Intrinsic Bias Metrics Do Not Correlate with Application Bias , 2020, ACL.
[24] Muhammad Abdul-Mageed,et al. ARBERT & MARBERT: Deep Bidirectional Transformers for Arabic , 2020, ACL.
[25] Benoit Sagot,et al. When Being Unseen from mBERT is just the Beginning: Handling New Languages With Multilingual Language Models , 2020, NAACL.
[26] Colin Raffel,et al. mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer , 2020, NAACL.
[27] Samuel R. Bowman,et al. CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models , 2020, EMNLP.
[28] Mascha Kurpicz-Briki,et al. Cultural Differences in Bias? Origin and Gender Bias in Pre-Trained German and French Word Embeddings , 2020, SwissText/KONVENS.
[29] Alan Ritter,et al. An Empirical Study of Pre-trained Transformers for Arabic Information Extraction , 2020, EMNLP.
[30] Siva Reddy,et al. StereoSet: Measuring stereotypical bias in pretrained language models , 2020, ACL.
[31] Orhan Firat,et al. XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating Cross-lingual Generalization , 2020, ICML.
[32] Andrei Barbu,et al. Measuring Social Biases in Grounded Vision and Language Embeddings , 2020, NAACL.
[33] Hazem M. Hajj,et al. AraBERT: Transformer-based Model for Arabic Language Understanding , 2020, OSACT.
[34] Myle Ott,et al. Unsupervised Cross-lingual Representation Learning at Scale , 2019, ACL.
[35] Nanyun Peng,et al. The Woman Worked as a Babysitter: On Biases in Language Generation , 2019, EMNLP.
[36] Benoît Sagot,et al. Asynchronous Pipeline for Processing Huge Corpora on Medium to Low Resource Infrastructures , 2019 .
[37] Jayadev Bhaskaran,et al. Good Secretaries, Bad Truck Drivers? Occupational Gender Stereotypes in Sentiment Analysis , 2019, Proceedings of the First Workshop on Gender Bias in Natural Language Processing.
[38] Goran Glavas,et al. Are We Consistently Biased? Multidimensional Analysis of Biases in Distributional Word Vectors , 2019, *SEMEVAL.
[39] Jieyu Zhao,et al. Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods , 2018, NAACL.
[40] Rachael Tatman,et al. Gender and Dialect Bias in YouTube’s Automatic Captions , 2017, EthNLP@EACL.
[41] Arvind Narayanan,et al. Semantics derived automatically from language corpora contain human-like biases , 2016, Science.
[42] Nayeon Lee,et al. Hate Speech Classifiers are Culturally Insensitive , 2023, C3NLP.
[43] Bryan C. Semaan,et al. Toward Cultural Bias Evaluation Datasets: The Case of Bengali Gender, Religious, and National Identity , 2023, C3NLP.
[44] Lilja Øvrelid,et al. Occupational Biases in Norwegian and Multilingual Language Models , 2022, GEBNLP.
[45] Karën Fort,et al. French CrowS-Pairs: Extending a challenge dataset for measuring social bias in masked language models to a language other than English , 2022, ACL.
[46] Dragomir R. Radev,et al. You reap what you sow: On the Challenges of Bias Evaluation Under Multilingual Settings , 2022, BIGSCIENCE.
[47] Federico Bianchi,et al. Measuring Harmful Sentence Completion in Language Models for LGBTQIA+ Individuals , 2022, LTEDI.
[48] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.