Explaining Algorithmic Fairness Through Fairness-Aware Causal Path Decomposition

Algorithmic fairness has aroused considerable interests in data mining and machine learning communities recently. So far the existing research has been mostly focusing on the development of quantitative metrics to measure algorithm disparities across different protected groups, and approaches for adjusting the algorithm output to reduce such disparities. In this paper, we propose to study the problem of identification of the source of model disparities. Unlike existing interpretation methods which typically learn feature importance, we consider the causal relationships among feature variables and propose a novel framework to decompose the disparity into the sum of contributions from fairness-aware causal paths, which are paths linking the sensitive attribute and the final predictions, on the graph. We also consider the scenario when the directions on certain edges within those paths cannot be determined. Our framework is also model agnostic and applicable to a variety of quantitative disparity measures. Empirical evaluations on both synthetic and real-world data sets are provided to show that our method can provide precise and comprehensive explanations to the model disparities.

[1]  A. Adegunsoye,et al.  Association of Black Race with Outcomes in COVID-19 Disease: A Retrospective Cohort Study , 2020, Annals of the American Thoracic Society.

[2]  Giorgos Borboudakis,et al.  Incorporating Causal Prior Knowledge as Path-Constraints in Bayesian Networks and Maximal Ancestral Graphs , 2012, ICML.

[3]  Xintao Wu,et al.  Causal Modeling-Based Discrimination Discovery and Removal: Criteria, Bounds, and Algorithms , 2019, IEEE Transactions on Knowledge and Data Engineering.

[4]  Scott Lundberg,et al.  A Unified Approach to Interpreting Model Predictions , 2017, NIPS.

[5]  Tom Burr,et al.  Causation, Prediction, and Search , 2003, Technometrics.

[6]  Krishna P. Gummadi,et al.  Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification without Disparate Mistreatment , 2016, WWW.

[7]  Martin T. Wells,et al.  Fairness criteria through the lens of directed acyclic graphical models , 2019, ArXiv.

[8]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[9]  Jenna Wiens,et al.  Shapley Flow: A Graph-based Approach to Interpreting Model Predictions , 2020, ArXiv.

[10]  J. Feldman,et al.  Plan and operation of the NHANES I Epidemiologic Followup Study, 1992. , 1997, Vital and health statistics. Ser. 1, Programs and collection procedures.

[11]  Aapo Hyvärinen,et al.  Nonlinear independent component analysis: Existence and uniqueness results , 1999, Neural Networks.

[12]  Julia A. Wolfson,et al.  Food Insecurity and COVID-19: Disparities in Early Effects for US Adults , 2020, Nutrients.

[13]  Ilya Feige,et al.  Explainability for fair machine learning , 2020, ArXiv.

[14]  Silvia Chiappa,et al.  Path-Specific Counterfactual Fairness , 2018, AAAI.

[15]  Colin Rowat,et al.  Asymmetric Shapley values: incorporating causal knowledge into model-agnostic explainability , 2020, NeurIPS.

[16]  Tianqi Chen,et al.  XGBoost: A Scalable Tree Boosting System , 2016, KDD.

[17]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[18]  Emilija Perkovi'c Identifying causal effects in maximally oriented partially directed acyclic graphs , 2019, UAI.

[19]  Hanghang Tong,et al.  PC-Fairness: A Unified Framework for Measuring Causality-based Fairness , 2019, NeurIPS.

[20]  Tom Claassen,et al.  Causal Shapley Values: Exploiting Causal Knowledge to Explain Individual Predictions of Complex Models , 2020, NeurIPS.

[21]  Christopher T. Lowenkamp,et al.  False Positives, False Negatives, and False Analyses: A Rejoinder to "Machine Bias: There's Software Used across the Country to Predict Future Criminals. and It's Biased against Blacks" , 2016 .

[22]  E. Price-Haywood,et al.  Hospitalization and Mortality among Black Patients and White Patients with Covid-19 , 2020, The New England journal of medicine.

[23]  Kjersti Aas,et al.  Explaining individual predictions when features are dependent: More accurate approximations to Shapley values , 2019, Artif. Intell..

[24]  Bernhard Schölkopf,et al.  Avoiding Discrimination through Causal Reasoning , 2017, NIPS.

[25]  Matt J. Kusner,et al.  Counterfactual Fairness , 2017, NIPS.

[26]  Toniann Pitassi,et al.  Learning Adversarially Fair and Transferable Representations , 2018, ICML.

[27]  E. Klein,et al.  SARS-CoV-2 Positivity Rate for Latinos in the Baltimore-Washington, DC Region. , 2020, JAMA.

[28]  Stefano Ermon,et al.  Learning Controllable Fair Representations , 2018, AISTATS.

[29]  Ilya Shpitser,et al.  Fair Inference on Outcomes , 2017, AAAI.

[30]  John Langford,et al.  A Reductions Approach to Fair Classification , 2018, ICML.

[31]  Blake Lemoine,et al.  Mitigating Unwanted Biases with Adversarial Learning , 2018, AIES.