Statistical inference for individual fairness

As we rely on machine learning (ML) models to make more consequential decisions, the issue of ML models perpetuating or even exacerbating undesirable historical biases (e.g. gender and racial biases) has come to the fore of the public’s attention. In this paper, we focus on the problem of detecting violations of individual fairness in ML models. We formalize the problem as measuring the susceptibility of ML models against a form of adversarial attack and develop a suite of inference tools for the adversarial cost function. The tools allow auditors to assess the individual fairness of ML models in a statistically-principled way: form confidence intervals for the worst-case performance differential between similar individuals and test hypotheses of model fairness with (asymptotic) non-coverage/Type I error rate control. We demonstrate the utility of our tools in a real-world case study1.

[1]  Michael T. Heath,et al.  Scientific Computing , 2018 .

[2]  John Langford,et al.  A Reductions Approach to Fair Classification , 2018, ICML.

[3]  Yaacov Ritov,et al.  On conditional parity as a notion of non-discrimination in machine learning , 2017, ArXiv.

[4]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[5]  John Duchi,et al.  Statistics of Robust Optimization: A Generalized Empirical Likelihood Approach , 2016, Math. Oper. Res..

[6]  Karthyek R. A. Murthy,et al.  Quantifying Distributional Model Risk Via Optimal Transport , 2016, Math. Oper. Res..

[7]  Jonathon Shlens,et al.  Explaining and Harnessing Adversarial Examples , 2014, ICLR.

[8]  Percy Liang,et al.  Fairness Without Demographics in Repeated Loss Minimization , 2018, ICML.

[9]  Michael T. Heath,et al.  Scientific Computing: An Introductory Survey , 1996 .

[10]  Yuekai Sun,et al.  Two Simple Ways to Learn Individual Fairness Metrics from Data , 2020, ICML.

[11]  M. Kearns,et al.  Fairness in Criminal Justice Risk Assessments: The State of the Art , 2017, Sociological Methods & Research.

[12]  Alexandra Chouldechova,et al.  What’s in a Name? Reducing Bias in Bios without Access to Protected Attributes , 2019, NAACL.

[13]  Krishna P. Gummadi,et al.  An Empirical Study on Learning Fairness Metrics for COMPAS Data with Human Supervision , 2019, ArXiv.

[14]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[15]  R. Shprintzen,et al.  What's in a name? , 1990, The Cleft palate journal.

[16]  Matt J. Kusner,et al.  Counterfactual Fairness , 2017, NIPS.

[17]  Christina Ilvento,et al.  Metric Learning for Individual Fairness , 2019, FORC.

[18]  Yuekai Sun,et al.  Auditing ML Models for Individual Bias and Unfairness , 2020, AISTATS.

[19]  Rachel K. E. Bellamy,et al.  AI Fairness 360: An Extensible Toolkit for Detecting, Understanding, and Mitigating Unwanted Algorithmic Bias , 2018, ArXiv.