Algorithmic Recourse: from Counterfactual Explanations to Interventions

As machine learning is increasingly used to inform consequential decision-making (e.g., pre-trial bail and loan approval), it becomes important to explain how the system arrived at its decision, and also suggest actions to achieve a favorable decision. Counterfactual explanations -"how the world would have (had) to be different for a desirable outcome to occur"- aim to satisfy these criteria. Existing works have primarily focused on designing algorithms to obtain counterfactual explanations for a wide range of settings. However, it has largely been overlooked that ultimately, one of the main objectives is to allow people to act rather than just understand. In layman's terms, counterfactual explanations inform an individual where they need to get to, but not how to get there. In this work, we rely on causal reasoning to caution against the use of counterfactual explanations as a recommendable set of actions for recourse. Instead, we propose a shift of paradigm from recourse via nearest counterfactual explanations to recourse through minimal interventions, shifting the focus from explanations to interventions.

[1]  Judea Pearl Causality by Judea Pearl , 2009 .

[2]  Mark Alfano,et al.  The philosophical basis of algorithmic recourse , 2020, FAT*.

[3]  Peter A. Flach,et al.  FACE: Feasible and Actionable Counterfactual Explanations , 2020, AIES.

[4]  Aws Albarghouthi,et al.  Synthesizing Action Sequences for Modifying Model Decisions , 2019, AAAI.

[5]  Emily Chen,et al.  How do Humans Understand Explanations from Machine Learning Systems? An Evaluation of the Human-Interpretability of Explanation , 2018, ArXiv.

[6]  Tim Miller,et al.  Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..

[7]  Solon Barocas,et al.  The hidden assumptions behind counterfactual explanations and principal reasons , 2019, FAT*.

[8]  Amit Sharma,et al.  Preserving Causal Constraints in Counterfactual Explanations for Machine Learning Classifiers , 2019, ArXiv.

[9]  Matt J. Kusner,et al.  When Worlds Collide: Integrating Different Counterfactual Assumptions in Fairness , 2017, NIPS.

[10]  Joichi Ito,et al.  Interventions over Predictions: Reframing the Ethical Debate for Actuarial Risk Assessment , 2017, FAT.

[11]  David Danks,et al.  Causal discovery algorithms: A practical guide , 2018 .

[12]  Kevin B. Korb,et al.  Varieties of Causal Intervention , 2004, PRICAI.

[13]  Cynthia Rudin,et al.  Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead , 2018, Nature Machine Intelligence.

[14]  Amit Dhurandhar,et al.  Explanations based on the Missing: Towards Contrastive Explanations with Pertinent Negatives , 2018, NeurIPS.

[15]  Cynthia Rudin,et al.  Please Stop Explaining Black Box Models for High Stakes Decisions , 2018, ArXiv.

[16]  Franco Turini,et al.  Local Rule-Based Explanations of Black Box Decision Systems , 2018, ArXiv.

[17]  Chandan Singh,et al.  Definitions, methods, and applications in interpretable machine learning , 2019, Proceedings of the National Academy of Sciences.

[18]  Joseph Y. Halpern,et al.  Abstracting Causal Models , 2018, AAAI.

[19]  Judea Pearl,et al.  A Probabilistic Calculus of Actions , 1994, UAI.

[20]  Amir-Hossein Karimi,et al.  Model-Agnostic Counterfactual Explanations for Consequential Decisions , 2019, AISTATS.

[21]  C. Allen,et al.  Stanford Encyclopedia of Philosophy , 2011 .

[22]  Bernhard Schölkopf,et al.  Avoiding Discrimination through Causal Reasoning , 2017, NIPS.

[23]  Matt J. Kusner,et al.  Counterfactual Fairness , 2017, NIPS.

[24]  Bernhard Schölkopf,et al.  Causal Consistency of Structural Equation Models , 2017, UAI.

[25]  Martin Wattenberg,et al.  The What-If Tool: Interactive Probing of Machine Learning Models , 2019, IEEE Transactions on Visualization and Computer Graphics.

[26]  Been Kim,et al.  Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.

[27]  Marie-Jeanne Lesot,et al.  Inverse Classification for Comparison-based Interpretability in Machine Learning , 2017, ArXiv.

[28]  Oluwasanmi Koyejo,et al.  Towards Realistic Individual Recourse and Actionable Explanations in Black-Box Decision Making Systems , 2019, ArXiv.

[29]  Leif Hancox-Li,et al.  Robustness in machine learning explanations: does it matter? , 2020, FAT*.

[30]  Frederick Eberhardt,et al.  Introduction to the foundations of causal discovery , 2017, International Journal of Data Science and Analytics.

[31]  P. Spirtes,et al.  Causation, prediction, and search , 1993 .

[32]  Silvia Chiappa,et al.  Path-Specific Counterfactual Fairness , 2018, AAAI.

[33]  Issa Kohler-Hausmann Eddie Murphy and the Dangers of Counterfactual Causal Thinking About Detecting Racial Discrimination , 2019 .

[34]  R. Scheines,et al.  Interventions and Causal Inference , 2007, Philosophy of Science.

[35]  Mathias Frisch,et al.  Causation and intervention , 2014 .

[36]  Chris Russell,et al.  Efficient Search for Diverse Coherent Explanations , 2019, FAT.

[37]  Zachary C. Lipton,et al.  The mythos of model interpretability , 2018, Commun. ACM.

[38]  David Gunning,et al.  DARPA's explainable artificial intelligence (XAI) program , 2019, IUI.

[39]  Chris Russell,et al.  Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR , 2017, ArXiv.

[40]  J. Pearl Causality: Models, Reasoning and Inference , 2000 .

[41]  Bernhard Schölkopf,et al.  Elements of Causal Inference: Foundations and Learning Algorithms , 2017 .

[42]  Luís Moniz Pereira,et al.  On Indicative Conditionals , 2015, IWOST-1.

[43]  J. Pearl,et al.  Causal Inference in Statistics: A Primer , 2016 .

[44]  P. Spirtes,et al.  Review of Causal Discovery Methods Based on Graphical Models , 2019, Front. Genet..

[45]  Donald Nute,et al.  Counterfactuals , 1975, Notre Dame J. Formal Log..

[46]  Judea Pearl,et al.  Structural Counterfactuals: A Brief Introduction , 2013, Cogn. Sci..

[47]  Amit Sharma,et al.  Explaining machine learning classifiers through diverse counterfactual explanations , 2020, FAT*.

[48]  Paul Voigt,et al.  The EU General Data Protection Regulation (GDPR) , 2017 .

[49]  Joydeep Ghosh,et al.  CERTIFAI: Counterfactual Explanations for Robustness, Transparency, Interpretability, and Fairness of Artificial Intelligence models , 2019, ArXiv.

[50]  Suresh Venkatasubramanian,et al.  Equalizing Recourse across Groups , 2019, ArXiv.

[51]  Ro'i Zultan,et al.  Causal Responsibility and Counterfactuals , 2013, Cogn. Sci..

[52]  Yang Liu,et al.  Actionable Recourse in Linear Classification , 2018, FAT.

[53]  Stefan Rüping,et al.  Learning interpretable models , 2006 .