Measuring the Biases that Matter: The Ethical and Casual Foundations for Measures of Fairness in Algorithms

Measures of algorithmic bias can be roughly classified into four categories, distinguished by the conditional probabilistic dependencies to which they are sensitive. First, measures of "procedural bias" diagnose bias when the score returned by an algorithm is probabilistically dependent on a sensitive class variable (e.g. race or sex). Second, measures of "outcome bias" capture probabilistic dependence between class variables and the outcome for each subject (e.g. parole granted or loan denied). Third, measures of "behavior-relative error bias" capture probabilistic dependence between class variables and the algorithmic score, conditional on target behaviors (e.g. recidivism or loan default). Fourth, measures of "score-relative error bias" capture probabilistic dependence between class variables and behavior, conditional on score. Several recent discussions have demonstrated a tradeoff between these different measures of algorithmic bias, and at least one recent paper has suggested conditions under which tradeoffs may be minimized. In this paper we use the machinery of causal graphical models to show that, under standard assumptions, the underlying causal relations among variables forces some tradeoffs. We delineate a number of normative considerations that are encoded in different measures of bias, with reference to the philosophical literature on the wrongfulness of disparate treatment and disparate impact. While both kinds of error bias are nominally motivated by concern to avoid disparate impact, we argue that consideration of causal structures shows that these measures are better understood as complicated and unreliable measures of procedural biases (i.e. disparate treatment). Moreover, while procedural bias is indicative of disparate treatment, we show that the measure of procedural bias one ought to adopt is dependent on the account of the wrongfulness of disparate treatment one endorses. Finally, given that neither score-relative nor behavior-relative measures of error bias capture the relevant normative considerations, we suggest that error bias proper is best measured by score-based measures of accuracy, such as the Brier score.

[1]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[2]  Carlos Eduardo Scheidegger,et al.  Certifying and Removing Disparate Impact , 2014, KDD.

[3]  Philip Pettit,et al.  Republicanism: A Theory of Freedom and Government , 1997 .

[4]  P. Spirtes,et al.  Causation, Prediction, and Search, 2nd Edition , 2001 .

[5]  M. Kearns,et al.  Fairness in Criminal Justice Risk Assessments: The State of the Art , 2017, Sociological Methods & Research.

[6]  Krishna P. Gummadi,et al.  Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification without Disparate Mistreatment , 2016, WWW.

[7]  J. Pearl Causality: Models, Reasoning and Inference , 2000 .

[8]  Bernard R. Boxill Blacks and social justice , 1984 .

[9]  Bernhard Schölkopf,et al.  Avoiding Discrimination through Causal Reasoning , 2017, NIPS.

[10]  Peter Spirtes,et al.  Introduction to Causal Inference , 2010, J. Mach. Learn. Res..

[11]  Alexandra Chouldechova,et al.  Fair prediction with disparate impact: A study of bias in recidivism prediction instruments , 2016, Big Data.

[12]  K. Lippert‐Rasmussen Indirect Discrimination is Not Necessarily Unjust , 2014 .

[13]  Elizabeth Anderson,et al.  What Is the Point of Equality? , 1999, Ethics.

[14]  Richard D. Kahlenberg The Remedy: Class, Race, And Affirmative Action , 1996 .

[15]  Matt J. Kusner,et al.  Counterfactual Fairness , 2017, NIPS.

[16]  Judea Pearl,et al.  The International Journal of Biostatistics C AUSAL I NFERENCE An Introduction to Causal Inference , 2011 .

[17]  Matt J. Kusner,et al.  Causal Interventions for Fairness , 2018, ArXiv.

[18]  K. Lippert‐Rasmussen The badness of discrimination , 2006 .

[19]  G. Brier VERIFICATION OF FORECASTS EXPRESSED IN TERMS OF PROBABILITY , 1950 .

[20]  T. Scanlon,et al.  Moral Dimensions , 2010 .

[21]  Alexandra Chouldechova,et al.  Does mitigating ML's impact disparity require treatment disparity? , 2017, NeurIPS.

[22]  Richard J. Arneson Discrimination, Disparate Impact, and Theories of Justice , 2013 .

[23]  Deborah Hellman,et al.  When Is Discrimination Wrong , 2008 .

[24]  Indre Zliobaite,et al.  On the relation between accuracy and fairness in binary classification , 2015, ArXiv.

[25]  Andrew D. Selbst,et al.  Big Data's Disparate Impact , 2016 .

[26]  Sharad Goel,et al.  The Measure and Mismeasure of Fairness: A Critical Review of Fair Machine Learning , 2018, ArXiv.

[27]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[28]  Avi Feller,et al.  Algorithmic Decision Making and the Cost of Fairness , 2017, KDD.

[29]  S. Segall What's so Bad about Discrimination? , 2012, Utilitas.

[30]  Patrick S. Shin THE SUBSTANTIVE PRINCIPLE OF EQUAL TREATMENT , 2009, Legal Theory.

[31]  Suresh Venkatasubramanian,et al.  On the (im)possibility of fairness , 2016, ArXiv.

[32]  Henry Chai Fairness in machine learning lecture notes (Lecture 25) , 2019 .

[33]  Jon M. Kleinberg,et al.  Inherent Trade-Offs in the Fair Determination of Risk Scores , 2016, ITCS.

[34]  Reuben Binns,et al.  Fairness in Machine Learning: Lessons from Political Philosophy , 2017, FAT.

[35]  Cass R. Sunstein,et al.  The Anticaste Principle , 1994 .

[36]  Vaughn R. McKim,et al.  Causality in crisis? : statistical methods and the search for causal knowledge in the social sciences , 1998 .