Discovering and Categorising Language Biases in Reddit

We present a data-driven approach using word embeddings to discover and categorise language biases on the discussion platform Reddit. As spaces for isolated user communities, platforms such as Reddit are increasingly connected to issues of racism, sexism and other forms of discrimination. Hence, there is a need to monitor the language of these groups. One of the most promising AI approaches to trace linguistic biases in large textual datasets involves word embeddings, which transform text into high-dimensional dense vectors and capture semantic relations between words. Yet, previous studies require predefined sets of potential biases to study, e.g., whether gender is more or less associated with particular types of jobs. This makes these approaches unfit to deal with smaller and community-centric datasets such as those on Reddit, which contain smaller vocabularies and slang, as well as biases that may be particular to that community. This paper proposes a data-driven approach to automatically discover language biases encoded in the vocabulary of online discourse communities on Reddit. In our approach, protected attributes are connected to evaluative words found in the data, which are then categorised through a semantic analysis system. We verify the effectiveness of our method by comparing the biases we discover in the Google News dataset with those found in previous literature. We then successfully discover gender bias, religion bias, and ethnic bias in different Reddit communities. We conclude by discussing potential application scenarios and limitations of this data-driven bias discovery method.

[1]  Camiel J. Beukeboom,et al.  Mechanisms of linguistic bias: How words reflect and maintain stereotypic expectancies , 2014 .

[2]  Solon Barocas,et al.  Roles for computing in social change , 2019, FAT*.

[3]  Saif Mohammad,et al.  Examining Gender and Race Bias in Two Hundred Sentiment Analysis Systems , 2018, *SEMEVAL.

[4]  Brian A. Nosek,et al.  Harvesting implicit group attitudes and beliefs from a demonstration web site , 2002 .

[5]  Cecilia Ovesdotter Alm,et al.  An Analysis of Domestic Abuse Discourse on Reddit , 2015, EMNLP.

[6]  Natalia Criado,et al.  Attesting Biases and Discrimination using Language Semantics , 2019, ArXiv.

[7]  Munmun De Choudhury,et al.  Detecting and Characterizing Mental Health Related Self-Disclosure in Social Media , 2015, CHI Extended Abstracts.

[8]  Virginia R. Dominguez Mapping the Language of Racism: Discourse and the Legitimation of Exploitation. MARGARET WETHERELL and JONATHAN POTTER , 1995 .

[9]  Jason Weston,et al.  Natural Language Processing (Almost) from Scratch , 2011, J. Mach. Learn. Res..

[10]  Alice E. Marwick,et al.  Media Manipulation and Disinformation Online , 2017 .

[11]  Della Summers,et al.  Longman Dictionary of Contemporary English , 1995 .

[12]  Ingmar Weber,et al.  Automated Hate Speech Detection and the Problem of Offensive Language , 2017, ICWSM.

[14]  Krishna P. Gummadi,et al.  Beyond Distributive Fairness in Algorithmic Decision Making: Feature Selection for Procedurally Fair Learning , 2018, AAAI.

[15]  J. B. Mountford,et al.  Topic Modeling The Red Pill , 2018 .

[16]  Paul Rayson,et al.  ASSIST: Automated Semantic Assistance for Translators , 2006, EACL.

[17]  Hinrich Schütze,et al.  Word Embedding Calculus in Meaningful Ultradense Subspaces , 2016, ACL.

[18]  David Mimno,et al.  Evaluating the Stability of Embedding-based Word Similarities , 2018, TACL.

[19]  Paul Rayson,et al.  Automatic Content Analysis of Spoken Discourse , 1992 .

[20]  Alan W Black,et al.  Measuring Bias in Contextualized Word Representations , 2019, Proceedings of the First Workshop on Gender Bias in Natural Language Processing.

[21]  Daniel Jurafsky,et al.  Word embeddings quantify 100 years of gender and ethnic stereotypes , 2017, Proceedings of the National Academy of Sciences.

[22]  Zizi Papacharissi Toward New Journalism(s) , 2015 .

[23]  Luciano Arcuri,et al.  Nomina sunt omina: on the inductive potential of nouns and adjectives in person perception. , 2008, Journal of personality and social psychology.

[24]  Danna Zhou,et al.  d. , 1934, Microbial pathogenesis.

[25]  Eric Gilbert,et al.  VADER: A Parsimonious Rule-Based Model for Sentiment Analysis of Social Media Text , 2014, ICWSM.

[26]  Adam Tauman Kalai,et al.  Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings , 2016, NIPS.

[27]  G. G. Stokes "J." , 1890, The New Yale Book of Quotations.

[28]  Arvind Narayanan,et al.  Semantics derived automatically from language corpora contain human-like biases , 2016, Science.

[29]  Sudeep Bhatia,et al.  The semantic representation of prejudice and stereotypes , 2017, Cognition.

[30]  Ian F. Haney-López Dog Whistle Politics: How Coded Racial Appeals Have Reinvented Racism and Wrecked the Middle Class , 2014 .

[31]  A. Greenwald,et al.  Measuring individual differences in implicit cognition: the implicit association test. , 1998, Journal of personality and social psychology.

[32]  Emiel van Miltenburg Stereotyping and Bias in the Flickr30K Dataset , 2016, ArXiv.

[33]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[34]  Bernie Hogan,et al.  Using Platform Signals for Distinguishing Discourses: The Case of Men's Rights and Men's Liberation on Reddit , 2019, ICWSM.

[35]  Magnus Sahlgren,et al.  The Distributional Hypothesis , 2008 .

[36]  J. Swales,et al.  The Concept of I Discourse Community , 2011 .

[37]  Jeremy Blackburn,et al.  The Pushshift Reddit Dataset , 2020, ICWSM.

[38]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.