Bringing the People Back In: Contesting Benchmark Machine Learning Datasets

In response to algorithmic unfairness embedded in sociotechnical systems, significant attention has been focused on the contents of machine learning datasets which have revealed biases towards white, cisgender, male, and Western data subjects. In contrast, comparatively less attention has been paid to the histories, values, and norms embedded in such datasets. In this work, we outline a research program - a genealogy of machine learning data - for investigating how and why these datasets have been created, what and whose values influence the choices of data to collect, the contextual and contingent conditions of their creation. We describe the ways in which benchmark datasets in machine learning operate as infrastructure and pose four research questions for these datasets. This interrogation forces us to "bring the people back in" by aiding us in understanding the labor embedded in dataset construction, and thereby presenting new avenues of contestation for other researchers encountering the data.

[1]  Karthik Dinakar,et al.  Studying up: reorienting the study of algorithmic fairness around issues of power , 2020, FAT*.

[2]  B. Larkin The Politics and Poetics of Infrastructure , 2013 .

[3]  Inioluwa Deborah Raji,et al.  Model Cards for Model Reporting , 2018, FAT.

[4]  S. E. Sachs,et al.  The algorithm at work? Explanation and repair in the enactment of similarity in art data , 2020, Information, Communication & Society.

[5]  Deirdre K. Mulligan,et al.  Shaping Our Tools: Contestability as a Means to Promote Responsible Algorithmic Decision Making in the Professions , 2019 .

[6]  Ruha Benjamin Race After Technology: Abolitionist Tools for the New Jim Code , 2019, Social Forces.

[7]  K. Baker,et al.  Toward Information Infrastructure Studies: Ways of Knowing in a Networked Environment , 2009 .

[8]  Nick Seaver,et al.  Captivating algorithms: Recommender systems as traps , 2018, Journal of Material Culture.

[9]  Timnit Gebru,et al.  Lessons from archives: strategies for collecting sociocultural data in machine learning , 2019, FAT*.

[10]  D. Fitch,et al.  Review of "Algorithms of oppression: how search engines reinforce racism," by Noble, S. U. (2018). New York, New York: NYU Press. , 2018, CDQR.

[11]  J. H. Fujimura The Molecular Biological Bandwagon in Cancer Research: Where Social Worlds Meet , 1988 .

[12]  R. D'amico Discipline and Punish: The Birth of the Prison , 1978, Telos.

[13]  L. Ciaccia,et al.  The Immortal Life of Henrietta Lacks , 2010, The Yale Journal of Biology and Medicine.

[14]  Hannah Lebovits Automating Inequality: How High-Tech Tools Profile, Police, and Punish the Poor , 2018, Public Integrity.

[15]  B. Schutt Sorting things out. , 1970, The American journal of nursing.

[16]  M. Six Silberman,et al.  Turkopticon: interrupting worker invisibility in amazon mechanical turk , 2013, CHI.

[17]  Shari Trewin,et al.  AI Fairness for People with Disabilities: Point of View , 2018, ArXiv.

[18]  Zeynep Tufekci,et al.  Big Questions for Social Media Big Data: Representativeness, Validity and Other Methodological Pitfalls , 2014, ICWSM.

[19]  Adam Tauman Kalai,et al.  Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings , 2016, NIPS.

[20]  Smitha Milli,et al.  Value-laden disciplinary shifts in machine learning , 2019, FAT*.

[21]  Solon Barocas,et al.  Problem Formulation and Fairness , 2019, FAT.

[22]  Diana E. Forsythe,et al.  Ethics and Politics of Studying Up in Technoscience , 1999 .

[23]  Michael S. Bernstein,et al.  We Are Dynamo: Overcoming Stalling and Friction in Collective Action for Crowd Workers , 2015, CHI.

[24]  B. Latour,et al.  Laboratory Life: The Construction of Scientific Facts , 1979 .

[25]  Benjamin Heinzerling,et al.  NLP's Clever Hans Moment has Arrived , 2020 .

[26]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[27]  Emily Denton,et al.  Social Biases in NLP Models as Barriers for Persons with Disabilities , 2020, ACL.

[28]  R. Kohler, Lords of the fly: Drosophila genetics and the experimental life. , 1995 .

[29]  R. Stuart Geiger,et al.  Garbage in, garbage out?: do machine learning application papers in social computing report where human-labeled training data comes from? , 2019, FAT*.

[30]  Mary L. Gray,et al.  Ghost Work: How to Stop Silicon Valley from Building a New Global Underclass , 2019 .

[31]  Caitlin Lustig,et al.  How We've Taught Algorithms to See Identity: Constructing Race and Gender in Image Databases for Facial Analysis , 2020, Proc. ACM Hum. Comput. Interact..

[32]  Wiebe E. Bijker,et al.  Science in action : how to follow scientists and engineers through society , 1989 .

[33]  L. Nader Up the Anthropologist: Perspectives Gained From Studying Up. , 1972 .