Are Explanations Helpful? A Comparative Study of the Effects of Explanations in AI-Assisted Decision-Making
暂无分享,去创建一个
Ming Yin | Xinru Wang | Xinru Wang | Ming Yin
[1] Dong Si,et al. Machine Learning Classification of Tree Cover Type and Application to Forest Management , 2018, Int. J. Multim. Data Eng. Manag..
[2] Carrie J. Cai,et al. The effects of example-based explanations in a machine learning interface , 2019, IUI.
[3] Dong Si,et al. Addressing Forest Management Challenges by Refining Tree Cover Type Classification with Machine Learning Models , 2017, 2017 IEEE International Conference on Information Reuse and Integration (IRI).
[4] Tim Miller,et al. Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..
[5] Scott Lundberg,et al. A Unified Approach to Interpreting Model Predictions , 2017, NIPS.
[6] D. Goldstein,et al. Simple Rules for Complex Decisions , 2017, 1702.04690.
[7] Han Liu,et al. "Why is 'Chicago' deceptive?" Towards Building Model-Driven Tutorials for Humans , 2020, CHI.
[8] Been Kim,et al. Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.
[9] Anind K. Dey,et al. Investigating intelligibility for uncertain context-aware applications , 2011, UbiComp '11.
[10] Oluwasanmi Koyejo,et al. Examples are not enough, learn to criticize! Criticism for Interpretability , 2016, NIPS.
[11] Emil Pitkin,et al. Peeking Inside the Black Box: Visualizing Statistical Learning With Plots of Individual Conditional Expectation , 2013, 1309.6392.
[12] Vivian Lai,et al. On Human Predictions with Explanations and Predictions of Machine Learning Models: A Case Study on Deception Detection , 2018, FAT.
[13] Alun D. Preece,et al. Interpretable to Whom? A Role-based Model for Analyzing Interpretable Machine Learning Systems , 2018, ArXiv.
[14] Deborah Lee,et al. Are Well-Calibrated Users Effective Users? Associations Between Calibration of Trust and Performance on an Automation-Aided Task , 2015, Hum. Factors.
[15] Martin Wattenberg,et al. Interpretability Beyond Feature Attribution: Quantitative Testing with Concept Activation Vectors (TCAV) , 2017, ICML.
[16] Daniel C. Murrie,et al. Do scores from risk measures matter to jurors , 2013 .
[17] Hanna M. Wallach,et al. A Human-Centered Agenda for Intelligible Machine Learning , 2021 .
[18] Eric Horvitz,et al. Beyond Accuracy: The Role of Mental Models in Human-AI Team Performance , 2019, HCOMP.
[19] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[20] Ruth M. J. Byrne,et al. Counterfactuals in Explainable Artificial Intelligence (XAI): Evidence from Human Reasoning , 2019, IJCAI.
[21] Emily Chen,et al. How do Humans Understand Explanations from Machine Learning Systems? An Evaluation of the Human-Interpretability of Explanation , 2018, ArXiv.
[22] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[23] Steve Whittaker,et al. Progressive disclosure: empirically motivated approaches to designing effective transparency , 2019, IUI.
[24] Jean Scholtz,et al. How do visual explanations foster end users' appropriate trust in machine learning? , 2020, IUI.
[25] Johannes Gehrke,et al. Intelligible Models for HealthCare: Predicting Pneumonia Risk and Hospital 30-day Readmission , 2015, KDD.
[26] Raymond Fok,et al. Does the Whole Exceed its Parts? The Effect of AI Explanations on Complementary Team Performance , 2020, CHI.
[27] Bonnie M. Muir,et al. Trust Between Humans and Machines, and the Design of Decision Aids , 1987, Int. J. Man Mach. Stud..
[28] Denis J. Dean,et al. Comparative accuracies of artificial neural networks and discriminant analysis in predicting forest cover types from cartographic variables , 1999 .
[29] Percy Liang,et al. Understanding Black-box Predictions via Influence Functions , 2017, ICML.
[30] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[31] Raja Parasuraman,et al. Humans and Automation: Use, Misuse, Disuse, Abuse , 1997, Hum. Factors.
[32] BEN GREEN,et al. The Principles and Limits of Algorithm-in-the-Loop Decision Making , 2019, Proc. ACM Hum. Comput. Interact..
[33] Brian Y. Lim,et al. COGAM: Measuring and Moderating Cognitive Load in Machine Learning Model Explanations , 2020, CHI.
[34] Anind K. Dey,et al. Why and why not explanations improve the intelligibility of context-aware intelligent systems , 2009, CHI.
[35] G. Cumming. The New Statistics: Why and How , 2013 .
[36] Krzysztof Z. Gajos,et al. Proxy tasks and subjective measures can be misleading in evaluating explainable AI systems , 2020, IUI.
[37] Jarke J. van Wijk,et al. Instance-Level Explanations for Fraud Detection: A Case Study , 2018, ICML 2018.
[38] Andrew Zisserman,et al. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.
[39] D. G. Robinson,et al. Danger Ahead: Risk Assessment and the Future of Bail Reform , 2018 .
[40] K. Heilbrun,et al. The Prediction of Criminal Recidivism in Juveniles , 2001 .
[41] Xia Hu,et al. Techniques for interpretable machine learning , 2018, Commun. ACM.
[42] M. McDaniel,et al. Make It Stick , 2014 .
[43] Qiaozhu Mei,et al. Feature-Based Explanations Don't Help People Detect Misclassifications of Online Toxicity , 2020, ICWSM.
[44] Enrico Costanza,et al. Evaluating saliency map explanations for convolutional neural networks: a user study , 2020, IUI.
[45] Nadine B. Sarter,et al. Supporting Trust Calibration and the Effective Use of Decision Aids by Presenting Dynamic System Confidence Information , 2006, Hum. Factors.
[46] Rachel K. E. Bellamy,et al. Explaining models an empirical study of how explanations impact fairness judgment , 2019 .
[47] Cynthia Rudin,et al. Falling Rule Lists , 2014, AISTATS.
[48] Haiyi Zhu,et al. Explaining Decision-Making Algorithms through UI: Strategies to Help Non-Expert Stakeholders , 2019, CHI.
[49] Ming Yin,et al. Understanding the Effect of Accuracy on Trust in Machine Learning Models , 2019, CHI.
[50] Jure Leskovec,et al. Interpretable Decision Sets: A Joint Framework for Description and Prediction , 2016, KDD.
[51] Chris Russell,et al. Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR , 2017, ArXiv.
[52] Steven M. Drucker,et al. Gamut: A Design Probe to Understand How Data Scientists Understand Machine Learning Models , 2019, CHI.
[53] K. Gummadi,et al. Human Decision Making with Machine Assistance , 2019, Proc. ACM Hum. Comput. Interact..
[54] Raja Parasuraman,et al. A Design Methodology for Trust Cue Calibration in Cognitive Agents , 2014, HCI.
[55] Cynthia Rudin,et al. Model Class Reliance: Variable Importance Measures for any Machine Learning Model Class, from the "Rashomon" Perspective , 2018 .
[56] Jun Zhao,et al. 'It's Reducing a Human Being to a Percentage': Perceptions of Justice in Algorithmic Decisions , 2018, CHI.
[57] Yunfeng Zhang,et al. Effect of confidence and explanation on accuracy and trust calibration in AI-assisted decision making , 2020, FAT*.
[58] Daniel G. Goldstein,et al. Manipulating and Measuring Model Interpretability , 2018, CHI.
[59] Barbara Oakley,et al. A Mind for Numbers: How to Excel at Math and Science (Even If You Flunked Algebra) , 2014 .
[60] N. McGlynn. Thinking fast and slow. , 2014, Australian veterinary journal.
[61] G. Cumming. Understanding the New Statistics: Effect Sizes, Confidence Intervals, and Meta-Analysis , 2011 .
[62] Amina Adadi,et al. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) , 2018, IEEE Access.
[63] Jacob Andreas,et al. Are Visual Explanations Useful? A Case Study in Model-in-the-Loop Prediction , 2020, ArXiv.
[64] Pierre Dragicevic,et al. Fair Statistical Communication in HCI , 2016 .
[65] Rich Caruana,et al. Model compression , 2006, KDD '06.
[66] Zachary Chase Lipton. The mythos of model interpretability , 2016, ACM Queue.