Problems with Shapley-value-based explanations as feature importance measures
暂无分享,去创建一个
[1] Daniel G. Goldstein,et al. Manipulating and Measuring Model Interpretability , 2018, CHI.
[2] Scott Lundberg,et al. A Unified Approach to Interpreting Model Predictions , 2017, NIPS.
[3] Mark Alfano,et al. The philosophical basis of algorithmic recourse , 2020, FAT*.
[4] Kjersti Aas,et al. Explaining individual predictions when features are dependent: More accurate approximations to Shapley values , 2019, Artif. Intell..
[5] W. Kruskal. Relative Importance by Averaging Over Orderings , 1987 .
[6] Sergio Araya. Algorithmic Transparency , 2008 .
[7] Ankur Taly,et al. Explainable machine learning in deployment , 2019, FAT*.
[8] Art B. Owen,et al. On Shapley Value for Measuring Importance of Dependent Inputs , 2016, SIAM/ASA J. Uncertain. Quantification.
[9] Suresh Venkatasubramanian,et al. Auditing black-box models for indirect influence , 2016, Knowledge and Information Systems.
[10] Colin Rowat,et al. Asymmetric Shapley values: incorporating causal knowledge into model-agnostic explainability , 2019, NeurIPS.
[11] Dale T. Miller,et al. Norm theory: Comparing reality to its alternatives , 1986 .
[12] Tim Miller,et al. Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..
[13] Solon Barocas,et al. The Intuitive Appeal of Explainable Machines , 2018 .
[14] Ariel Rubinstein,et al. A Course in Game Theory , 1995 .
[15] Harmanpreet Kaur,et al. Interpreting Interpretability: Understanding Data Scientists' Use of Interpretability Tools for Machine Learning , 2020, CHI.
[16] Been Kim,et al. Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.
[17] S. Lipovetsky,et al. Analysis of regression in game theory approach , 2001 .
[18] Mykola Pechenizkiy,et al. A Human-Grounded Evaluation of SHAP for Alert Processing , 2019, ArXiv.
[19] Yair Zick,et al. Algorithmic Transparency via Quantitative Input Influence: Theory and Experiments with Learning Systems , 2016, 2016 IEEE Symposium on Security and Privacy (SP).
[20] Yang Liu,et al. Actionable Recourse in Linear Classification , 2018, FAT.
[21] Giles Hooker,et al. Please Stop Permuting Features: An Explanation and Alternatives , 2019, ArXiv.
[22] Ankur Taly,et al. The Explanation Game: Explaining Machine Learning Models Using Shapley Values , 2020, CD-MAKE.
[23] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[24] Erik Strumbelj,et al. Explaining prediction models and individual predictions with feature contributions , 2014, Knowledge and Information Systems.
[25] Sameer Singh,et al. Fooling LIME and SHAP: Adversarial Attacks on Post hoc Explanation Methods , 2020, AIES.
[26] Solon Barocas,et al. The hidden assumptions behind counterfactual explanations and principal reasons , 2019, FAT*.
[27] Scott M. Lundberg,et al. Consistent Individualized Feature Attribution for Tree Ensembles , 2018, ArXiv.
[28] Ankur Taly,et al. The Explanation Game: Explaining Machine Learning Models with Cooperative Game Theory , 2019, ArXiv.
[29] Chris Russell,et al. Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR , 2017, ArXiv.