Having Beer after Prayer? Measuring Cultural Bias in Large Language Models

Are language models culturally biased? It is important that language models conform to the cultural aspects of the communities they serve. However, we show in this paper that language models suffer from a significant bias towards Western culture when handling and generating text in Arabic, often preferring, and producing Western-fitting content as opposed to the relevant Arab content. We quantify this bias through a likelihood scoring-based metric using naturally occurring contexts that we collect from online social media. Our experiments reveal that both Arabic monolingual and multilingual models exhibit bias towards Western culture in eight different cultural aspects: person names, food, clothing, location, literature, beverage, religion, and sports. Models also tend to exhibit more bias when prompted with Arabic sentences that are more linguistically aligned with English. These findings raise concerns about the cultural relevance of current language models. Our analyses show that providing culture-indicating tokens or culturally-relevant demonstrations to the model can help in debiasing.

[1]  Daniel Hershcovich,et al.  Assessing Cross-Cultural Alignment between ChatGPT and Human Societies: An Empirical Study , 2023, C3NLP.

[2]  Parishad BehnamGhader,et al.  An Analysis of Social Biases Present in BERT Variants Across Multiple Languages , 2022, ArXiv.

[3]  Alexander M. Rush,et al.  BLOOM: A 176B-Parameter Open-Access Multilingual Language Model , 2022, ArXiv.

[4]  Dan Jurafsky,et al.  Multilingual BERT has an accent: Evaluating English influences on fluency in multilingual models , 2022, FINDINGS.

[5]  Partha P. Talukdar,et al.  Re-contextualizing Fairness in NLP: The Case of India , 2022, AACL.

[6]  Yang Trista Cao,et al.  Theory-Grounded Measurement of U.S. Social Stereotypes in English Language Models , 2022, NAACL.

[7]  Kathleen C. Fraser,et al.  Does Moral Code have a Moral Code? Probing Delphi’s Moral Philosophy , 2022, TRUSTNLP.

[8]  Eric Michael Smith,et al.  “I’m sorry to hear that”: Finding New Biases in Language Models with a Holistic Descriptor Dataset , 2022, EMNLP.

[9]  Naoaki Okazaki,et al.  Gender Bias in Masked Language Models for Multiple Languages , 2022, NAACL.

[10]  Tatiana Shavrina,et al.  mGPT: Few-Shot Learners Go Multilingual , 2022, ArXiv.

[11]  Yadollah Yaghoobzadeh,et al.  Metaphors in Pre-Trained Language Models: Probing and Generalization Across Datasets and Languages , 2022, ACL.

[12]  Noah A. Smith,et al.  Annotators with Attitudes: How Annotator Beliefs And Identities Bias Toxic Language Detection , 2021, NAACL.

[13]  Alice H. Oh,et al.  Mitigating Language-Dependent Ethnic Bias in BERT , 2021, EMNLP.

[14]  Kashif Shah,et al.  Quantifying Social Biases in NLP: A Generalization and Empirical Comparison of Extrinsic Fairness Metrics , 2021, Transactions of the Association for Computational Linguistics.

[15]  Dirk Hovy,et al.  HONEST: Measuring Hurtful Sentence Completion in Language Models , 2021, NAACL.

[16]  Abubakar Abid,et al.  Large language models associate Muslims with violence , 2021, Nature Machine Intelligence.

[17]  C. Rothkopf,et al.  Large pre-trained language models contain human-like biases of what is right and wrong to do , 2021, Nature Machine Intelligence.

[18]  Arkaitz Zubiaga,et al.  Towards generalisable hate speech detection: a review on obstacles and solutions , 2021, PeerJ Comput. Sci..

[19]  Yejin Choi,et al.  Challenges in Automated Debiasing for Toxic Language Detection , 2021, EACL.

[20]  James Zou,et al.  Persistent Anti-Muslim Bias in Large Language Models , 2021, AIES.

[21]  Hazem M. Hajj,et al.  AraGPT2: Pre-Trained Transformer for Arabic Language Generation , 2020, WANLP.

[22]  Iryna Gurevych,et al.  UNKs Everywhere: Adapting Multilingual Language Models to New Scripts , 2020, EMNLP.

[23]  Adam Lopez,et al.  Intrinsic Bias Metrics Do Not Correlate with Application Bias , 2020, ACL.

[24]  Muhammad Abdul-Mageed,et al.  ARBERT & MARBERT: Deep Bidirectional Transformers for Arabic , 2020, ACL.

[25]  Benoit Sagot,et al.  When Being Unseen from mBERT is just the Beginning: Handling New Languages With Multilingual Language Models , 2020, NAACL.

[26]  Colin Raffel,et al.  mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer , 2020, NAACL.

[27]  Samuel R. Bowman,et al.  CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models , 2020, EMNLP.

[28]  Mascha Kurpicz-Briki,et al.  Cultural Differences in Bias? Origin and Gender Bias in Pre-Trained German and French Word Embeddings , 2020, SwissText/KONVENS.

[29]  Alan Ritter,et al.  An Empirical Study of Pre-trained Transformers for Arabic Information Extraction , 2020, EMNLP.

[30]  Siva Reddy,et al.  StereoSet: Measuring stereotypical bias in pretrained language models , 2020, ACL.

[31]  Orhan Firat,et al.  XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating Cross-lingual Generalization , 2020, ICML.

[32]  Andrei Barbu,et al.  Measuring Social Biases in Grounded Vision and Language Embeddings , 2020, NAACL.

[33]  Hazem M. Hajj,et al.  AraBERT: Transformer-based Model for Arabic Language Understanding , 2020, OSACT.

[34]  Myle Ott,et al.  Unsupervised Cross-lingual Representation Learning at Scale , 2019, ACL.

[35]  Nanyun Peng,et al.  The Woman Worked as a Babysitter: On Biases in Language Generation , 2019, EMNLP.

[36]  Benoît Sagot,et al.  Asynchronous Pipeline for Processing Huge Corpora on Medium to Low Resource Infrastructures , 2019 .

[37]  Jayadev Bhaskaran,et al.  Good Secretaries, Bad Truck Drivers? Occupational Gender Stereotypes in Sentiment Analysis , 2019, Proceedings of the First Workshop on Gender Bias in Natural Language Processing.

[38]  Goran Glavas,et al.  Are We Consistently Biased? Multidimensional Analysis of Biases in Distributional Word Vectors , 2019, *SEMEVAL.

[39]  Jieyu Zhao,et al.  Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods , 2018, NAACL.

[40]  Rachael Tatman,et al.  Gender and Dialect Bias in YouTube’s Automatic Captions , 2017, EthNLP@EACL.

[41]  Arvind Narayanan,et al.  Semantics derived automatically from language corpora contain human-like biases , 2016, Science.

[42]  Nayeon Lee,et al.  Hate Speech Classifiers are Culturally Insensitive , 2023, C3NLP.

[43]  Bryan C. Semaan,et al.  Toward Cultural Bias Evaluation Datasets: The Case of Bengali Gender, Religious, and National Identity , 2023, C3NLP.

[44]  Lilja Øvrelid,et al.  Occupational Biases in Norwegian and Multilingual Language Models , 2022, GEBNLP.

[45]  Karën Fort,et al.  French CrowS-Pairs: Extending a challenge dataset for measuring social bias in masked language models to a language other than English , 2022, ACL.

[46]  Dragomir R. Radev,et al.  You reap what you sow: On the Challenges of Bias Evaluation Under Multilingual Settings , 2022, BIGSCIENCE.

[47]  Federico Bianchi,et al.  Measuring Harmful Sentence Completion in Language Models for LGBTQIA+ Individuals , 2022, LTEDI.

[48]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.