Algorithmic Recourse: from Counterfactual Explanations to Interventions

As machine learning is increasingly used to inform consequential decision-making (e.g., pre-trial bail and loan approval), it becomes important to explain how the system arrived at its decision, and also suggest actions to achieve a favorable decision. Counterfactual explanations -- "how the world would have (had) to be different for a desirable outcome to occur" -- aim to satisfy these criteria. Existing works have primarily focused on designing algorithms to obtain counterfactual explanations for a wide range of settings. However, one of the main objectives of "explanations as a means to help a data-subject act rather than merely understand" has been overlooked. In layman's terms, counterfactual explanations inform an individual where they need to get to, but not how to get there. In this work, we rely on causal reasoning to caution against the use of counterfactual explanations as a recommendable set of actions for recourse. Instead, we propose a shift of paradigm from recourse via nearest counterfactual explanations to recourse through minimal interventions, moving the focus from explanations to recommendations. Finally, we provide the reader with an extensive discussion on how to realistically achieve recourse beyond structural interventions.

[1]  Marie-Jeanne Lesot,et al.  Inverse Classification for Comparison-based Interpretability in Machine Learning , 2017, ArXiv.

[2]  Oluwasanmi Koyejo,et al.  Towards Realistic Individual Recourse and Actionable Explanations in Black-Box Decision Making Systems , 2019, ArXiv.

[3]  Martin Wattenberg,et al.  The What-If Tool: Interactive Probing of Machine Learning Models , 2019, IEEE Transactions on Visualization and Computer Graphics.

[4]  Ro'i Zultan,et al.  Causal Responsibility and Counterfactuals , 2013, Cogn. Sci..

[5]  C. Allen,et al.  Stanford Encyclopedia of Philosophy , 2011 .

[6]  Emily Chen,et al.  How do Humans Understand Explanations from Machine Learning Systems? An Evaluation of the Human-Interpretability of Explanation , 2018, ArXiv.

[7]  Frederick Eberhardt,et al.  Introduction to the foundations of causal discovery , 2017, International Journal of Data Science and Analytics.

[8]  R. Scheines,et al.  Interventions and Causal Inference , 2007, Philosophy of Science.

[9]  Solon Barocas,et al.  The hidden assumptions behind counterfactual explanations and principal reasons , 2019, FAT*.

[10]  Amit Sharma,et al.  Preserving Causal Constraints in Counterfactual Explanations for Machine Learning Classifiers , 2019, ArXiv.

[11]  Matt J. Kusner,et al.  Counterfactual Fairness , 2017, NIPS.

[12]  Paul Voigt,et al.  The EU General Data Protection Regulation (GDPR) , 2017 .

[13]  Franco Turini,et al.  Local Rule-Based Explanations of Black Box Decision Systems , 2018, ArXiv.

[14]  Joseph Y. Halpern,et al.  Abstracting Causal Models , 2018, AAAI.

[15]  Matt J. Kusner,et al.  When Worlds Collide: Integrating Different Counterfactual Assumptions in Fairness , 2017, NIPS.

[16]  Stefan Rüping,et al.  Learning interpretable models , 2006 .

[17]  Amit Sharma,et al.  Explaining machine learning classifiers through diverse counterfactual explanations , 2020, FAT*.

[18]  Aws Albarghouthi,et al.  Synthesizing Action Sequences for Modifying Model Decisions , 2019, AAAI.

[19]  Bernhard Schölkopf,et al.  Elements of Causal Inference: Foundations and Learning Algorithms , 2017 .

[20]  Suresh Venkatasubramanian,et al.  Equalizing Recourse across Groups , 2019, ArXiv.

[21]  Joichi Ito,et al.  Interventions over Predictions: Reframing the Ethical Debate for Actuarial Risk Assessment , 2017, FAT.

[22]  Amir-Hossein Karimi,et al.  Model-Agnostic Counterfactual Explanations for Consequential Decisions , 2019, AISTATS.

[23]  Mathias Frisch,et al.  Causation and intervention , 2014 .

[24]  Been Kim,et al.  Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.

[25]  Joydeep Ghosh,et al.  CERTIFAI: Counterfactual Explanations for Robustness, Transparency, Interpretability, and Fairness of Artificial Intelligence models , 2019, ArXiv.

[26]  P. Spirtes,et al.  Review of Causal Discovery Methods Based on Graphical Models , 2019, Front. Genet..

[27]  Cynthia Rudin,et al.  Please Stop Explaining Black Box Models for High Stakes Decisions , 2018, ArXiv.

[28]  Issa Kohler-Hausmann Eddie Murphy and the Dangers of Counterfactual Causal Thinking About Detecting Racial Discrimination , 2019 .

[29]  Silvia Chiappa,et al.  Path-Specific Counterfactual Fairness , 2018, AAAI.

[30]  Peter A. Flach,et al.  FACE: Feasible and Actionable Counterfactual Explanations , 2020, AIES.

[31]  Amit Dhurandhar,et al.  Explanations based on the Missing: Towards Contrastive Explanations with Pertinent Negatives , 2018, NeurIPS.

[32]  Bernhard Schölkopf,et al.  Avoiding Discrimination through Causal Reasoning , 2017, NIPS.

[33]  Chris Russell,et al.  Efficient Search for Diverse Coherent Explanations , 2019, FAT.

[34]  Yang Liu,et al.  Actionable Recourse in Linear Classification , 2018, FAT.

[35]  Kevin B. Korb,et al.  Varieties of Causal Intervention , 2004, PRICAI.

[36]  Chandan Singh,et al.  Definitions, methods, and applications in interpretable machine learning , 2019, Proceedings of the National Academy of Sciences.

[37]  J. Pearl Causality: Models, Reasoning and Inference , 2000 .

[38]  Luís Moniz Pereira,et al.  On Indicative Conditionals , 2015, IWOST-1.

[39]  Judea Pearl,et al.  Structural Counterfactuals: A Brief Introduction , 2013, Cogn. Sci..

[40]  Tim Miller,et al.  Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..

[41]  Judea Pearl,et al.  A Probabilistic Calculus of Actions , 1994, UAI.

[42]  David Gunning,et al.  DARPA's explainable artificial intelligence (XAI) program , 2019, IUI.

[43]  Leif Hancox-Li,et al.  Robustness in machine learning explanations: does it matter? , 2020, FAT*.

[44]  David Danks,et al.  Causal discovery algorithms: A practical guide , 2018 .

[45]  Cynthia Rudin,et al.  Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead , 2018, Nature Machine Intelligence.

[46]  Chris Russell,et al.  Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR , 2017, ArXiv.

[47]  Bernhard Schölkopf,et al.  Causal Consistency of Structural Equation Models , 2017, UAI.

[48]  Zachary Chase Lipton The mythos of model interpretability , 2016, ACM Queue.

[49]  Tom Burr,et al.  Causation, Prediction, and Search , 2003, Technometrics.

[50]  J. Pearl,et al.  Causal Inference in Statistics: A Primer , 2016 .

[51]  Mark Alfano,et al.  The philosophical basis of algorithmic recourse , 2020, FAT*.