Evaluating XAI: A comparison of rule-based and example-based explanations

Abstract Current developments in Artificial Intelligence (AI) led to a resurgence of Explainable AI (XAI). New methods are being researched to obtain information from AI systems in order to generate explanations for their output. However, there is an overall lack of valid and reliable evaluations of the effects on users' experience of, and behavior in response to explanations. New XAI methods are often based on an intuitive notion what an effective explanation should be. Rule- and example-based contrastive explanations are two exemplary explanation styles. In this study we evaluate the effects of these two explanation styles on system understanding, persuasive power and task performance in the context of decision support in diabetes self-management. Furthermore, we provide three sets of recommendations based on our experience designing this evaluation to help improve future evaluations. Our results show that rule-based explanations have a small positive effect on system understanding, whereas both rule- and example-based explanations seem to persuade users in following the advice even when incorrect. Neither explanation improves task performance compared to no explanation. This can be explained by the fact that both explanation styles only provide details relevant for a single decision, not the underlying rational or causality. These results show the importance of user evaluations in assessing the current assumptions and intuitions on effective explanations.

[1]  Monika Reddy,et al.  Type 1 diabetes in adults: supporting self management , 2016, British Medical Journal.

[2]  Maartje M. A. de Graaf,et al.  How People Explain Action (and Autonomous Intelligent Systems Should Too) , 2017, AAAI Fall Symposia.

[3]  Trevor Darrell,et al.  Generating Visual Explanations , 2016, ECCV.

[4]  Dympna O'Sullivan,et al.  The Role of Explanations on Trust and Reliance in Clinical Decision Support Systems , 2015, 2015 International Conference on Healthcare Informatics.

[5]  Weng-Keen Wong,et al.  Principles of Explanatory Debugging to Personalize Interactive Machine Learning , 2015, IUI.

[6]  K. Branting,et al.  Building Explanations from Rules and Structured Cases , 1991, Int. J. Man Mach. Stud..

[7]  Mark A. Neerincx,et al.  Towards a Theory of Longitudinal Trust Calibration in Human–Robot Teams , 2019, International Journal of Social Robotics.

[8]  M. Maruthappu,et al.  Artificial intelligence in diabetes care , 2018, Diabetic medicine : a journal of the British Diabetic Association.

[9]  Subbarao Kambhampati,et al.  Explicability? Legibility? Predictability? Transparency? Privacy? Security? The Emerging Landscape of Interpretable Agent Behavior , 2018, ICAPS.

[10]  Gwenn Englebienne,et al.  How model accuracy and explanation fidelity influence user trust , 2019, IJCAI 2019.

[11]  Or Biran,et al.  Explanation and Justification in Machine Learning : A Survey Or , 2017 .

[12]  Georg Langs,et al.  Causability and explainability of artificial intelligence in medicine , 2019, WIREs Data Mining Knowl. Discov..

[13]  Matthew W. Lewis,et al.  Self-Explonations: How Students Study and Use Examples in Learning to Solve Problems , 1989, Cogn. Sci..

[14]  R. Tibshirani,et al.  Prototype selection for interpretable classification , 2011, 1202.5933.

[15]  R. Atkinson Optimizing learning from examples using animated pedagogical agents. , 2002 .

[16]  Gary Klein,et al.  Metrics for Explainable AI: Challenges and Prospects , 2018, ArXiv.

[17]  Riccardo Satta,et al.  LEAFAGE: Example-based and Feature importance-based Explanations for Black-box ML models , 2018, 2019 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE).

[18]  Cynthia Rudin,et al.  Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead , 2018, Nature Machine Intelligence.

[19]  Franco Turini,et al.  A Survey of Methods for Explaining Black Box Models , 2018, ACM Comput. Surv..

[20]  Orit Zaslavsky,et al.  Counter-Examples That (Only) Prove and Counter-Examples That (Also) Explain. , 1997 .

[21]  Fikri Gökpinar,et al.  A SIMULATION STUDY ON TESTS FOR ONE-WAY ANOVA UNDER THE UNEQUAL VARIANCE ASSUMPTION , 2010 .

[22]  C. J. Huberty,et al.  Multivariate analysis versus multiple univariate analyses. , 1989 .

[23]  Vibhu O. Mittal,et al.  Generating explanations in context: The system perspective , 1995 .

[24]  Amina Adadi,et al.  Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) , 2018, IEEE Access.

[25]  Bradley Hayes,et al.  Improving Robot Controller Transparency Through Autonomous Policy Explanation , 2017, 2017 12th ACM/IEEE International Conference on Human-Robot Interaction (HRI.

[26]  Pamela J. Hinds,et al.  Autonomy and Common Ground in Human-Robot Interaction: A Field Study , 2007, IEEE Intelligent Systems.

[27]  A. Renkl Worked-out examples: instructional explanations support learning by self- explanations , 2002 .

[28]  John Riedl,et al.  Explaining collaborative filtering recommendations , 2000, CSCW '00.

[29]  Joachim Diederich,et al.  Eclectic Rule-Extraction from Support Vector Machines , 2005 .

[30]  J. Kirk,et al.  Reliability and Validity in Qualitative Research , 1985 .

[31]  Cynthia Rudin,et al.  Falling Rule Lists , 2014, AISTATS.

[32]  Anind K. Dey,et al.  Assessing demand for intelligibility in context-aware applications , 2009, UbiComp.

[33]  Panagiotis G. Ipeirotis,et al.  Running Experiments on Amazon Mechanical Turk , 2010, Judgment and Decision Making.

[34]  Dan Conway,et al.  How to Recommend?: User Trust Factors in Movie Recommender Systems , 2017, IUI.

[35]  J. Vehí,et al.  Artificial Intelligence for Diabetes Management and Decision Support: Literature Review , 2018, Journal of medical Internet research.

[36]  G. Keppel,et al.  Design and Analysis: A Researcher's Handbook , 1976 .

[37]  Thomas G. Dietterich,et al.  Interacting meaningfully with machine learning systems: Three experiments , 2009, Int. J. Hum. Comput. Stud..

[38]  Raymond J. Mooney,et al.  Explaining Recommendations: Satisfaction vs. Promotion , 2005 .

[39]  Michael J. Pazzani,et al.  Representation of electronic mail filtering profiles: a user study , 2000, IUI '00.

[40]  Patients’ Knowledge of Diabetes Mellitus in a Nigerian City , 2011 .

[41]  Ellen A. Drost,et al.  Validity and Reliability in Social Science Research. , 2011 .

[42]  Davide Calvaresi,et al.  Explainable Agents and Robots: Results from a Systematic Literature Review , 2019, AAMAS.

[43]  Cynthia Rudin,et al.  The Bayesian Case Model: A Generative Approach for Case-Based Reasoning and Prototype Classification , 2014, NIPS.

[44]  Kuangyan Song,et al.  "Why Should You Trust My Explanation?" Understanding Uncertainty in LIME Explanations , 2019 .

[45]  J. Pearl Causal inference in statistics: An overview , 2009 .

[46]  Tim Miller,et al.  Contrastive explanation: a structural-model approach , 2018, The Knowledge Engineering Review.

[47]  Sameer Singh,et al.  How can we fool LIME and SHAP? Adversarial Attacks on Post hoc Explanation Methods , 2019, ArXiv.

[48]  Zachary Chase Lipton The mythos of model interpretability , 2016, ACM Queue.

[49]  Been Kim,et al.  Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.

[50]  Yang Wang,et al.  Effects of Influence on User Trust in Predictive Decision Making , 2019, CHI Extended Abstracts.

[51]  Oluwasanmi Koyejo,et al.  Examples are not enough, learn to criticize! Criticism for Interpretability , 2016, NIPS.

[52]  I. Vlahavas,et al.  Machine Learning and Data Mining Methods in Diabetes Research , 2017, Computational and structural biotechnology journal.

[53]  Michael A. Rupp,et al.  Intelligent Agent Transparency in Human–Agent Teaming for Multi-UxV Management , 2016, Hum. Factors.

[54]  L. Richard Ye,et al.  The Impact of Explanation Facilities in User Acceptance of Expert System Advice , 1995, MIS Q..

[55]  Simone Stumpf,et al.  User Trust in Intelligent Systems: A Journey Over Time , 2016, IUI.

[56]  Tim Miller,et al.  Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..

[57]  Yiannis Demiris,et al.  Socio-Cognitive Engineering of a Robotic Partner for Child's Diabetes Self-Management , 2019, Front. Robot. AI.

[58]  Mark O. Riedl,et al.  Rationalization: A Neural Machine Translation Approach to Generating Natural Language Explanations , 2017, AIES.

[59]  Nancy J. Cooke,et al.  Designing a Synthetic Task Environment , 2017 .

[60]  Wolfgang Minker,et al.  The Impact of Explanation Dialogues on Human-Computer Trust , 2013, HCI.

[61]  Anind K. Dey,et al.  Why and why not explanations improve the intelligibility of context-aware intelligent systems , 2009, CHI.

[62]  Emily Chen,et al.  How do Humans Understand Explanations from Machine Learning Systems? An Evaluation of the Human-Interpretability of Explanation , 2018, ArXiv.

[63]  Mark A. Neerincx,et al.  Contrastive Explanations with Local Foil Trees , 2018, ICML 2018.