Multi-Dimensional Gender Bias Classification

Machine learning models are trained to find patterns in data. NLP models can inadvertently learn socially undesirable patterns when training on gender biased text. In this work, we propose a general framework that decomposes gender bias in text along several pragmatic and semantic dimensions: bias from the gender of the person being spoken about, bias from the gender of the person being spoken to, and bias from the gender of the speaker. Using this fine-grained framework, we automatically annotate eight large scale datasets with gender information. In addition, we collect a novel, crowdsourced evaluation benchmark of utterance-level gender rewrites. Distinguishing between gender bias along multiple dimensions is important, as it enables us to train finer-grained gender bias classifiers. We show our classifiers prove valuable for a variety of important applications, such as controlling for gender bias in generative models, detecting gender bias in arbitrary text, and shed light on offensive language in terms of genderedness.

[1]  David Bamman,et al.  Unsupervised Discovery of Biographical Structure from Text , 2014, TACL.

[2]  Liliana Simón,et al.  Ways of speaking , 2021, Language and Nationality.

[3]  Kim-Kwang Raymond Choo,et al.  User profiling in intrusion detection: A review , 2016, J. Netw. Comput. Appl..

[4]  Adam Tauman Kalai,et al.  Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings , 2016, NIPS.

[5]  Shlomo Argamon,et al.  Automatically Categorizing Written Texts by Author Gender , 2002, Lit. Linguistic Comput..

[6]  Mounia Lalmas,et al.  First Women, Second Sex: Gender Bias in Wikipedia , 2015, HT.

[7]  Yoshua Bengio,et al.  Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.

[8]  Yejin Choi,et al.  The Risk of Racial Bias in Hate Speech Detection , 2019, ACL.

[9]  Jieyu Zhao,et al.  Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods , 2018, NAACL.

[10]  Vicente Ordonez,et al.  Bias and Fairness in Natural Language Processing , 2019, EMNLP/IJCNLP.

[11]  Mai ElSherief,et al.  Towards Understanding Gender Bias in Relation Extraction , 2019, ACL.

[12]  Sara Mills,et al.  Language and Gender: Interdisciplinary Perspectives , 2014 .

[13]  Benedikt Hell,et al.  Comparison of Different Test Construction Strategies in the Development of a Gender Fair Interest Inventory Using Verbs , 2012 .

[14]  Noe Casas,et al.  Evaluating the Underlying Gender Bias in Contextualized Word Embeddings , 2019, Proceedings of the First Workshop on Gender Bias in Natural Language Processing.

[15]  Jason Weston,et al.  Wizard of Wikipedia: Knowledge-Powered Conversational agents , 2018, ICLR.

[16]  Guillaume Lample,et al.  Multiple-Attribute Text Style Transfer , 2018, ArXiv.

[17]  Chandler May,et al.  Social Bias in Elicited Natural Language Inferences , 2017, EthNLP@EACL.

[18]  Samuel R. Bowman,et al.  CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models , 2020, EMNLP.

[19]  Jackie Chi Kit Cheung,et al.  The KnowRef Coreference Corpus: Removing Gender and Number Cues for Difficult Pronominal Anaphora Resolution , 2018, ACL.

[20]  Dorothy A. Winsor,et al.  Talking Power: The Politics of Language , 1991 .

[21]  Jessika Weiss Women Men And Language A Sociolinguistic Account Of Gender Differences In Language , 2016 .

[22]  Alan W Black,et al.  Measuring Bias in Contextualized Word Representations , 2019, Proceedings of the First Workshop on Gender Bias in Natural Language Processing.

[23]  Daniel Jurafsky,et al.  Word embeddings quantify 100 years of gender and ethnic stereotypes , 2017, Proceedings of the National Academy of Sciences.

[24]  Nayeon Lee,et al.  Exploring Social Bias in Chatbots using Stereotype Knowledge , 2019, WNLP@ACL.

[25]  Noah A. Smith,et al.  Evaluating Gender Bias in Machine Translation , 2019, ACL.

[26]  Ryan Cotterell,et al.  Gender Bias in Contextualized Word Embeddings , 2019, NAACL.

[27]  Sima Sharifirad,et al.  Learning and Understanding Different Categories of Sexism Using Convolutional Neural Network’s Filters , 2019, WNLP@ACL.

[28]  Chandler May,et al.  On Measuring Social Biases in Sentence Encoders , 2019, NAACL.

[29]  Marta R. Costa-jussà,et al.  An analysis of gender bias studies in natural language processing , 2019, Nature Machine Intelligence.

[30]  Jason Weston,et al.  Queens are Powerful too: Mitigating Gender Bias in Dialogue Generation , 2020, EMNLP.

[31]  Yann Dauphin,et al.  Hierarchical Neural Story Generation , 2018, ACL.

[32]  f. bianchi Can You Translate that into Man? Commercial Machine Translation Systems Include Stylistic Biases , 2020 .

[33]  R. Lakoff Language and woman's place , 1973, Language in Society.

[34]  Haiyi Zhu,et al.  Monitoring the Gender Gap with Wikidata Human Gender Indicators , 2016, OpenSym.

[35]  Mai ElSherief,et al.  Mitigating Gender Bias in Natural Language Processing: Literature Review , 2019, ACL.

[36]  Ryan Cotterell,et al.  Examining Gender Bias in Languages with Grammatical Gender , 2019, EMNLP.

[37]  Stefania Degaetano-Ortlieb,et al.  Stylistic variation over 200 years of court proceedings according to gender and social class , 2018 .

[38]  Rajarathnam Chandramouli,et al.  Author gender identification from text , 2011, Digit. Investig..

[39]  Efstathios Stamatatos,et al.  A survey of modern authorship attribution methods , 2009, J. Assoc. Inf. Sci. Technol..

[40]  Yoav Goldberg,et al.  Lipstick on a Pig: Debiasing Methods Cover up Systematic Gender Biases in Word Embeddings But do not Remove Them , 2019, NAACL-HLT.

[41]  Max Klein,et al.  Wikipedia in the world of global gender inequality indices: what the biography gender gap is measuring , 2015, OpenSym.

[42]  Arvind Narayanan,et al.  Semantics derived automatically from language corpora contain human-like biases , 2016, Science.

[43]  Yejin Choi,et al.  Gender Attribution: Tracing Stylometric Evidence Beyond Topic and Genre , 2011, CoNLL.

[44]  C. Higgins,et al.  Gender differences in work-family conflict. , 1991 .

[45]  Joelle Pineau,et al.  The Second Conversational Intelligence Challenge (ConvAI2) , 2019, The NeurIPS '18 Competition.

[46]  Joan Swann,et al.  Girls, Boys, And Language , 1992 .

[47]  B Guerin Gender bias in the abstractness of verbs and adjectives. , 1994, The Journal of social psychology.

[48]  Aaron C. Kay,et al.  Journal of Personality and Social Psychology Evidence That Gendered Wording in Job Advertisements Exists and Sustains Gender Inequality , 2011 .

[49]  Yusu Qian,et al.  Reducing Gender Bias in Word-Level Language Models with a Gender-Equalizing Loss Function , 2019, ACL.

[50]  Adriana Kovashka,et al.  Authorship Attribution Using Probabilistic Context-Free Grammars , 2010, ACL.

[51]  Myron Wish,et al.  Perceived Dimensions of Interpersonal Relations , 1981 .

[52]  Ann Weatherall,et al.  Gender, Language and Discourse , 2002 .

[53]  Alyson Byrne,et al.  When She Brings Home the Job Status: Wives' Job Status, Status Leakage, and Marital Instability , 2017, Organ. Sci..

[54]  Stan Matwin,et al.  Using Attention-based Bidirectional LSTM to Identify Different Categories of Offensive Language Directed Toward Female Celebrities , 2019, WNLP@ACL.

[55]  Joseph M. Reagle,et al.  Gender Bias in Wikipedia and Britannica , 2011 .

[56]  Yoav Goldberg,et al.  How does Grammatical Gender Affect Noun Representations in Gender-Marking Languages? , 2019, CoNLL.

[57]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[58]  Rosa María Jiménez Catalán,et al.  Semantic derogation in animal metaphor: a contrastive-cognitive analysis of two male/female examples in English and Spanish , 2003 .

[59]  Anke A. Ehrhardt,et al.  Man and Woman, Boy and Girl: Differentiation and Dimorphism of Gender Identity from Conception to Maturity , 1973 .

[60]  David García,et al.  It's a Man's Wikipedia? Assessing Gender Inequality in an Online Encyclopedia , 2015, ICWSM.

[61]  A. Davis,et al.  Women, race & class , 1982 .

[62]  Ryan Cotterell,et al.  Unsupervised Discovery of Gendered Language through Latent-Variable Modeling , 2019, ACL.

[63]  Yusu Qian,et al.  Gender Stereotypes Differ between Male and Female Writings , 2019, ACL.

[64]  Frederick Mosteller,et al.  Applied Bayesian and classical inference : the case of the Federalist papers , 1984 .

[65]  Erik R. Thom,et al.  University of Pennsylvania Working Papers in Linguistics , 2007 .

[66]  Deborah Cameron,et al.  The Feminist Critique of Language: A Reader , 1990 .

[67]  Eduard Hovy,et al.  Generating Natural Language Under Pragmatic Constraints , 1988 .

[68]  E. Reid,et al.  Straying from breadwinning: Status and money in men's interpretations of their wives' work arrangements , 2018, Gender, Work & Organization.

[69]  Efstathios Stamatatos,et al.  Authorship Attribution for Social Media Forensics , 2017, IEEE Transactions on Information Forensics and Security.

[70]  Timnit Gebru,et al.  Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification , 2018, FAT.

[71]  H. T. Eddy The characteristic curves of composition. , 1887, Science.

[72]  Rosamund Moon From gorgeous to grumpy: adjectives, age, and gender , 2013 .

[73]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[74]  K. Crenshaw Demarginalizing the Intersection of Race and Sex: A Black Feminist Critique of Antidiscrimination Doctrine, Feminist Theory and Antiracist Politics , 1989 .

[75]  Allan Bell,et al.  Towards a sociolinguistics of style , 1997 .

[76]  J. Coates Women, Men, and Language: A Sociolinguistic Account of Sex Differences in Language , 1986 .

[77]  Zeyu Li,et al.  Learning Gender-Neutral Word Embeddings , 2018, EMNLP.

[78]  Jiliang Tang,et al.  Does Gender Matter? Towards Fairness in Dialogue Systems , 2020, COLING.

[79]  Ryan Cotterell,et al.  Counterfactual Data Augmentation for Mitigating Gender Stereotypes in Languages with Rich Morphology , 2019, ACL.

[80]  Solon Barocas,et al.  Language (Technology) is Power: A Critical Survey of “Bias” in NLP , 2020, ACL.

[81]  Haley Lepp Pardon the Interruption: Automatic Analysis of Gender and Competitive Turn-Taking in United States Supreme Court Hearings , 2019, WNLP@ACL.

[82]  Jörg Tiedemann,et al.  OpenSubtitles2016: Extracting Large Parallel Corpora from Movie and TV Subtitles , 2016, LREC.

[83]  Dirk Hovy,et al.  “You Sound Just Like Your Father” Commercial Machine Translation Systems Include Stylistic Biases , 2020, ACL.

[84]  Christine Charyton,et al.  Engineers' and musicians' choices of self-descriptive adjectives as potential indicators of creativity by gender and domain. , 2007 .

[85]  Yuanbin Wu,et al.  Exploring Human Gender Stereotypes with Word Association Test , 2019, EMNLP/IJCNLP.

[86]  A. Bell Language style as audience design , 1984, Language in Society.

[87]  G. Lakoff,et al.  Metaphors We Live by , 1982 .

[88]  P. Eckert,et al.  Language and Gender: Introduction to the study of language and gender , 2013 .

[89]  Lena Jaeger,et al.  Talking Difference On Gender And Language , 2016 .

[90]  Walter Daelemans,et al.  Authorship Attribution and Verification with Many Authors and Limited Data , 2008, COLING.

[91]  Rachael Tatman,et al.  Gender and Dialect Bias in YouTube’s Automatic Captions , 2017, EthNLP@EACL.

[92]  Alfredo Maldonado,et al.  Measuring Gender Bias in Word Embeddings across Domains and Discovering New Gender Bias Word Categories , 2019, Proceedings of the First Workshop on Gender Bias in Natural Language Processing.

[93]  G. Elder,et al.  WOMEN BEHIND THE MEN: , 1993 .

[94]  Carolyn E. Psenka,et al.  Exploring the Color of Glass: Letters of Recommendation for Female and Male Medical Faculty , 2003 .

[95]  Efstathios Stamatatos,et al.  Automatic Authorship Attribution , 1999, EACL.

[96]  Heather Savigny,et al.  Metaphors on Women in Academia: A Review of the Literature, 2004–2013 , 2015 .

[97]  Marta R. Costa-jussà,et al.  Equalizing Gender Bias in Neural Machine Translation with Word Embeddings Techniques , 2019, Proceedings of the First Workshop on Gender Bias in Natural Language Processing.

[98]  Dell Hymes,et al.  Explorations in the Ethnography of Speaking: Ways of Speaking , 1989 .

[99]  Yoav Goldberg,et al.  Null It Out: Guarding Protected Attributes by Iterative Nullspace Projection , 2020, ACL.

[100]  A. D. Shveĭt︠s︡er,et al.  Introduction to sociolinguistics , 1986 .

[101]  W. W. Philliber,et al.  Predicting Marital and Career Success among Dual-Worker Couples , 1982 .

[102]  Yejin Choi,et al.  Social Bias Frames: Reasoning about Social and Power Implications of Language , 2020, ACL.

[103]  Jason Weston,et al.  Learning to Speak and Act in a Fantasy Text Adventure Game , 2019, EMNLP.

[104]  Jason Weston,et al.  ParlAI: A Dialog Research Software Platform , 2017, EMNLP.

[105]  Rada Mihalcea,et al.  Women’s Syntactic Resilience and Men’s Grammatical Luck: Gender-Bias in Part-of-Speech Tagging and Dependency Parsing , 2019, ACL.

[106]  Corina Koolen,et al.  These are not the Stereotypes You are Looking For: Bias and Fairness in Authorial Gender Attribution , 2017, EthNLP@EACL.

[107]  D. Khor,et al.  “Doing Gender” , 2007 .

[108]  Angela Fan,et al.  Controllable Abstractive Summarization , 2017, NMT@ACL.

[109]  Graeme Hirst,et al.  Understanding Undesirable Word Embedding Associations , 2019, ACL.

[110]  Jason Weston,et al.  What makes a good conversation? How controllable attributes affect human judgments , 2019, NAACL.

[111]  Mahzarin R. Banaji,et al.  The Self in Social Contexts , 1994 .

[112]  Ryan Cotterell,et al.  It’s All in the Name: Mitigating Gender Bias with Name-Based Counterfactual Data Substitution , 2019, EMNLP.

[113]  Jane Sunderland,et al.  Language and Gender: An Advanced Resource Book , 2006 .

[114]  Danushka Bollegala,et al.  Gender-preserving Debiasing for Pre-trained Word Embeddings , 2019, ACL.

[115]  P. Eckert,et al.  Style and Sociolinguistic Variation. , 2002 .

[116]  Jason Weston,et al.  Poly-encoders: Transformer Architectures and Pre-training Strategies for Fast and Accurate Multi-sentence Scoring , 2019 .

[117]  Eduard Hovy,et al.  (Male, Bachelor) and (Female, Ph.D) have different connotations: Parallelly Annotated Stylistic Language Dataset with Multiple Personas , 2019, EMNLP.

[118]  Eduardo Graells-Garrido,et al.  Women through the glass ceiling: gender asymmetries in Wikipedia , 2016, EPJ Data Science.

[119]  Vicente Ordonez,et al.  Double-Hard Debias: Tailoring Word Embeddings for Gender Bias Mitigation , 2020, ACL.

[120]  Saif Mohammad,et al.  SemEval-2012 Task 2: Measuring Degrees of Relational Similarity , 2012, *SEMEVAL.

[121]  Stefanos Gritzalis,et al.  Effective identification of source code authors using byte-level information , 2006, ICSE.

[122]  Shlomo Argamon,et al.  Automatically profiling the author of an anonymous text , 2009, CACM.

[123]  Yang Trista Cao,et al.  Toward Gender-Inclusive Coreference Resolution , 2019, ACL.

[124]  Efstathios Stamatatos,et al.  Authorship Attribution Using Text Distortion , 2017, EACL.

[125]  Dirk Hovy,et al.  The Social Impact of Natural Language Processing , 2016, ACL.

[126]  Jason Weston,et al.  Build it Break it Fix it for Dialogue Safety: Robustness from Adversarial Human Attack , 2019, EMNLP.