Semantics derived automatically from language corpora contain human-like biases

Machines learn what people know implicitly AlphaGo has demonstrated that a machine can learn how to do things that people spend many years of concentrated study learning, and it can rapidly learn how to do them better than any human can. Caliskan et al. now show that machines can learn word associations from written texts and that these associations mirror those learned by humans, as measured by the Implicit Association Test (IAT) (see the Perspective by Greenwald). Why does this matter? Because the IAT has predictive value in uncovering the association between concepts, such as pleasantness and flowers or unpleasantness and insects. It can also tease out attitudes and beliefs—for example, associations between female names and family or male names and career. Such biases may not be expressed explicitly, yet they can prove influential in behavior. Science, this issue p. 183; see also p. 133 Computers can learn which words go together more or less often and can thus mimic human performance on a test of implicit bias. Machine learning is a means to derive artificial intelligence by discovering patterns in existing data. Here, we show that applying machine learning to ordinary human language results in human-like semantic biases. We replicated a spectrum of known biases, as measured by the Implicit Association Test, using a widely used, purely statistical machine-learning model trained on a standard corpus of text from the World Wide Web. Our results indicate that text corpora contain recoverable and accurate imprints of our historic biases, whether morally neutral as toward insects or flowers, problematic as toward race or gender, or even simply veridical, reflecting the status quo distribution of gender with respect to careers or first names. Our methods hold promise for identifying and addressing sources of bias in culture, including technology.

[1]  A. M. Turing,et al.  Computing Machinery and Intelligence , 1950, The Philosophy of Artificial Intelligence.

[2]  V.W. Zue,et al.  The use of speech knowledge in automatic speech recognition , 1985, Proceedings of the IEEE.

[3]  W. Marslen-Wilson,et al.  Accessing Different Types of Lexical Semantic Information: Evidence From Priming , 1995 .

[4]  M. Stubbs Text and corpus analysis , 1996 .

[5]  M. Stubbs Text and Corpus Analysis: Computer-Assisted Studies of Language and Culture , 1996 .

[6]  Will Lowe,et al.  Meaning and the Mental Lexicon , 1997, IJCAI.

[7]  A. Greenwald,et al.  Measuring individual differences in implicit cognition: the implicit association test. , 1998, Journal of personality and social psychology.

[8]  W. Lowe,et al.  Modelling functional priming and the associative boost , 1998 .

[9]  W. Lowe,et al.  The Direct Route: Mediated Priming in Semantic Space , 2000 .

[10]  Yoshua Bengio,et al.  A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..

[11]  Brian A. Nosek,et al.  Math Male , Me Female , Therefore Math Me , 2002 .

[12]  Brian A. Nosek,et al.  Math = male, me = female, therefore math ≠ me. , 2002 .

[13]  Brian A. Nosek,et al.  Harvesting implicit group attitudes and beliefs from a demonstration web site , 2002 .

[14]  宁北芳,et al.  疟原虫var基因转换速率变化导致抗原变异[英]/Paul H, Robert P, Christodoulou Z, et al//Proc Natl Acad Sci U S A , 2005 .

[15]  Yevgeniy B. Sirotin,et al.  Temporal associations and prior-list intrusions in free recall. , 2006, Journal of experimental psychology. Learning, memory, and cognition.

[16]  A. Kiefer,et al.  Implicit stereotypes and women's math performance : How implicit gender-math stereotypes influence women's susceptibility to stereotype threat , 2007 .

[17]  Cecilia,et al.  Are Emily and Greg More Employable Than Lakisha and Jamal ? A Field Experiment on Labor Market Discrimination , 2007 .

[18]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.

[19]  Kristinn R. Thórisson,et al.  Integrated A.I. systems , 2007, Minds and Machines.

[20]  Katherine D. Kinzler,et al.  The native language of social cognition , 2007, Proceedings of the National Academy of Sciences.

[21]  J. Bullinaria,et al.  Extracting semantic representations from word co-occurrence statistics: A computational study , 2007, Behavior research methods.

[22]  Magnus Sahlgren,et al.  The Distributional Hypothesis , 2008 .

[23]  Lawrence W Barsalou,et al.  Simulation, situated conceptualization, and prediction , 2009, Philosophical Transactions of the Royal Society B: Biological Sciences.

[24]  Brian A. Nosek,et al.  National differences in gender–science stereotypes predict national sex differences in science and math achievement , 2009, Proceedings of the National Academy of Sciences.

[25]  Patrick Pantel,et al.  From Frequency to Meaning: Vector Space Models of Semantics , 2010, J. Artif. Intell. Res..

[26]  Adrienne B. Dessel Prejudice in Schools: Promotion of an Inclusive Culture and Climate , 2010 .

[27]  M. Banaji,et al.  Implicit race attitudes predict trustworthiness judgments and economic trust decisions , 2011, Proceedings of the National Academy of Sciences.

[28]  W. Jeremy,et al.  Implicit and Explicit Stigmatizing Attitudes and Stereotypes About Depression , 2011 .

[29]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[30]  Jeffrey Dean,et al.  Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.

[31]  Toniann Pitassi,et al.  Learning Fair Representations , 2013, ICML.

[32]  T. Macfarlane Extracting Semantics from the Enron Corpus , 2013 .

[33]  Latanya Sweeney,et al.  Discrimination in online ad delivery , 2013, CACM.

[34]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[35]  A. Greenwald,et al.  With malice toward none and charity for some: ingroup favoritism enables discrimination. , 2014, The American psychologist.

[36]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[37]  Carlos Eduardo Scheidegger,et al.  Certifying and Removing Disparate Impact , 2014, KDD.

[38]  Antonio S Silva,et al.  Inter-Group Conflict and Cooperation: Field Experiments Before, During and After Sectarian Riots in Northern Ireland , 2015, Front. Psychol..

[39]  Andrew D. Selbst,et al.  Big Data's Disparate Impact , 2016 .

[40]  Daniel J. Lee Racial bias and the validity of the Implicit Association Test , 2016 .

[41]  K. Crawford Artificial Intelligence's White Guy Problem , 2016 .

[42]  M. Oswald,et al.  Norman Stanley Fletcher and the case of the proprietary algorithmic risk assessment. , 2016 .

[43]  Christopher T. Lowenkamp,et al.  False Positives, False Negatives, and False Analyses: A Rejoinder to "Machine Bias: There's Software Used across the Country to Predict Future Criminals. and It's Biased against Blacks" , 2016 .

[44]  G. Lupyan The Centrality of Language in Human Cognition , 2016 .

[45]  David G. Rand,et al.  Intuition, deliberation, and the evolution of cooperation , 2016, Proceedings of the National Academy of Sciences.

[46]  Braden A. Purcell,et al.  Hierarchical decision processes that operate over distinct timescales underlie choice and changes in strategy , 2016, Proceedings of the National Academy of Sciences.

[47]  Adam Tauman Kalai,et al.  Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings , 2016, NIPS.

[48]  L. Winner DO ARTIFACTS HAVE (cid:1) POLITICS? , 2022 .

[49]  Marc Hanheide,et al.  Robot task planning and explanation in open and uncertain worlds , 2017, Artif. Intell..

[50]  L. Christophorou Science , 2018, Emerging Dynamics: Science, Energy, Society and Values.