Evaluating XAI: A comparison of rule-based and example-based explanations
暂无分享,去创建一个
Jasper van der Waa | Anita Cremers | Elisabeth Nieuwburg | Mark A. Neerincx | J. V. D. Waa | A. Cremers | Mark Antonius Neerincx | Elisabeth G. I. Nieuwburg | Elisabeth Nieuwburg
[1] Monika Reddy,et al. Type 1 diabetes in adults: supporting self management , 2016, British Medical Journal.
[2] Maartje M. A. de Graaf,et al. How People Explain Action (and Autonomous Intelligent Systems Should Too) , 2017, AAAI Fall Symposia.
[3] Trevor Darrell,et al. Generating Visual Explanations , 2016, ECCV.
[4] Dympna O'Sullivan,et al. The Role of Explanations on Trust and Reliance in Clinical Decision Support Systems , 2015, 2015 International Conference on Healthcare Informatics.
[5] Weng-Keen Wong,et al. Principles of Explanatory Debugging to Personalize Interactive Machine Learning , 2015, IUI.
[6] K. Branting,et al. Building Explanations from Rules and Structured Cases , 1991, Int. J. Man Mach. Stud..
[7] Mark A. Neerincx,et al. Towards a Theory of Longitudinal Trust Calibration in Human–Robot Teams , 2019, International Journal of Social Robotics.
[8] M. Maruthappu,et al. Artificial intelligence in diabetes care , 2018, Diabetic medicine : a journal of the British Diabetic Association.
[9] Subbarao Kambhampati,et al. Explicability? Legibility? Predictability? Transparency? Privacy? Security? The Emerging Landscape of Interpretable Agent Behavior , 2018, ICAPS.
[10] Gwenn Englebienne,et al. How model accuracy and explanation fidelity influence user trust , 2019, IJCAI 2019.
[11] Or Biran,et al. Explanation and Justification in Machine Learning : A Survey Or , 2017 .
[12] Georg Langs,et al. Causability and explainability of artificial intelligence in medicine , 2019, WIREs Data Mining Knowl. Discov..
[13] Matthew W. Lewis,et al. Self-Explonations: How Students Study and Use Examples in Learning to Solve Problems , 1989, Cogn. Sci..
[14] R. Tibshirani,et al. Prototype selection for interpretable classification , 2011, 1202.5933.
[15] R. Atkinson. Optimizing learning from examples using animated pedagogical agents. , 2002 .
[16] Gary Klein,et al. Metrics for Explainable AI: Challenges and Prospects , 2018, ArXiv.
[17] Riccardo Satta,et al. LEAFAGE: Example-based and Feature importance-based Explanations for Black-box ML models , 2018, 2019 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE).
[18] Cynthia Rudin,et al. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead , 2018, Nature Machine Intelligence.
[19] Franco Turini,et al. A Survey of Methods for Explaining Black Box Models , 2018, ACM Comput. Surv..
[20] Orit Zaslavsky,et al. Counter-Examples That (Only) Prove and Counter-Examples That (Also) Explain. , 1997 .
[21] Fikri Gökpinar,et al. A SIMULATION STUDY ON TESTS FOR ONE-WAY ANOVA UNDER THE UNEQUAL VARIANCE ASSUMPTION , 2010 .
[22] C. J. Huberty,et al. Multivariate analysis versus multiple univariate analyses. , 1989 .
[23] Vibhu O. Mittal,et al. Generating explanations in context: The system perspective , 1995 .
[24] Amina Adadi,et al. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) , 2018, IEEE Access.
[25] Bradley Hayes,et al. Improving Robot Controller Transparency Through Autonomous Policy Explanation , 2017, 2017 12th ACM/IEEE International Conference on Human-Robot Interaction (HRI.
[26] Pamela J. Hinds,et al. Autonomy and Common Ground in Human-Robot Interaction: A Field Study , 2007, IEEE Intelligent Systems.
[27] A. Renkl. Worked-out examples: instructional explanations support learning by self- explanations , 2002 .
[28] John Riedl,et al. Explaining collaborative filtering recommendations , 2000, CSCW '00.
[29] Joachim Diederich,et al. Eclectic Rule-Extraction from Support Vector Machines , 2005 .
[30] J. Kirk,et al. Reliability and Validity in Qualitative Research , 1985 .
[31] Cynthia Rudin,et al. Falling Rule Lists , 2014, AISTATS.
[32] Anind K. Dey,et al. Assessing demand for intelligibility in context-aware applications , 2009, UbiComp.
[33] Panagiotis G. Ipeirotis,et al. Running Experiments on Amazon Mechanical Turk , 2010, Judgment and Decision Making.
[34] Dan Conway,et al. How to Recommend?: User Trust Factors in Movie Recommender Systems , 2017, IUI.
[35] J. Vehí,et al. Artificial Intelligence for Diabetes Management and Decision Support: Literature Review , 2018, Journal of medical Internet research.
[36] G. Keppel,et al. Design and Analysis: A Researcher's Handbook , 1976 .
[37] Thomas G. Dietterich,et al. Interacting meaningfully with machine learning systems: Three experiments , 2009, Int. J. Hum. Comput. Stud..
[38] Raymond J. Mooney,et al. Explaining Recommendations: Satisfaction vs. Promotion , 2005 .
[39] Michael J. Pazzani,et al. Representation of electronic mail filtering profiles: a user study , 2000, IUI '00.
[40] Patients’ Knowledge of Diabetes Mellitus in a Nigerian City , 2011 .
[41] Ellen A. Drost,et al. Validity and Reliability in Social Science Research. , 2011 .
[42] Davide Calvaresi,et al. Explainable Agents and Robots: Results from a Systematic Literature Review , 2019, AAMAS.
[43] Cynthia Rudin,et al. The Bayesian Case Model: A Generative Approach for Case-Based Reasoning and Prototype Classification , 2014, NIPS.
[44] Kuangyan Song,et al. "Why Should You Trust My Explanation?" Understanding Uncertainty in LIME Explanations , 2019 .
[45] J. Pearl. Causal inference in statistics: An overview , 2009 .
[46] Tim Miller,et al. Contrastive explanation: a structural-model approach , 2018, The Knowledge Engineering Review.
[47] Sameer Singh,et al. How can we fool LIME and SHAP? Adversarial Attacks on Post hoc Explanation Methods , 2019, ArXiv.
[48] Zachary Chase Lipton. The mythos of model interpretability , 2016, ACM Queue.
[49] Been Kim,et al. Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.
[50] Yang Wang,et al. Effects of Influence on User Trust in Predictive Decision Making , 2019, CHI Extended Abstracts.
[51] Oluwasanmi Koyejo,et al. Examples are not enough, learn to criticize! Criticism for Interpretability , 2016, NIPS.
[52] I. Vlahavas,et al. Machine Learning and Data Mining Methods in Diabetes Research , 2017, Computational and structural biotechnology journal.
[53] Michael A. Rupp,et al. Intelligent Agent Transparency in Human–Agent Teaming for Multi-UxV Management , 2016, Hum. Factors.
[54] L. Richard Ye,et al. The Impact of Explanation Facilities in User Acceptance of Expert System Advice , 1995, MIS Q..
[55] Simone Stumpf,et al. User Trust in Intelligent Systems: A Journey Over Time , 2016, IUI.
[56] Tim Miller,et al. Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..
[57] Yiannis Demiris,et al. Socio-Cognitive Engineering of a Robotic Partner for Child's Diabetes Self-Management , 2019, Front. Robot. AI.
[58] Mark O. Riedl,et al. Rationalization: A Neural Machine Translation Approach to Generating Natural Language Explanations , 2017, AIES.
[59] Nancy J. Cooke,et al. Designing a Synthetic Task Environment , 2017 .
[60] Wolfgang Minker,et al. The Impact of Explanation Dialogues on Human-Computer Trust , 2013, HCI.
[61] Anind K. Dey,et al. Why and why not explanations improve the intelligibility of context-aware intelligent systems , 2009, CHI.
[62] Emily Chen,et al. How do Humans Understand Explanations from Machine Learning Systems? An Evaluation of the Human-Interpretability of Explanation , 2018, ArXiv.
[63] Mark A. Neerincx,et al. Contrastive Explanations with Local Foil Trees , 2018, ICML 2018.