暂无分享,去创建一个
Carlos Eduardo Scheidegger | Sorelle A. Friedler | Dylan Slack | Chitradeep Dutta Roy | C. Scheidegger | Dylan Slack
[1] Solon Barocas,et al. The Intuitive Appeal of Explainable Machines , 2018 .
[2] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[3] Jure Leskovec,et al. Interpretable Decision Sets: A Joint Framework for Description and Prediction , 2016, KDD.
[4] Samuel J. Gershman,et al. Human-in-the-Loop Interpretability Prior , 2018, NeurIPS.
[5] S. M. García,et al. 2014: , 2020, A Party for Lazarus.
[6] Michael Veale,et al. Fairness and Accountability Design Needs for Algorithmic Support in High-Stakes Public Sector Decision-Making , 2018, CHI.
[7] Panagiotis Papapetrou,et al. A peek into the black box: exploring classifiers by randomization , 2014, Data Mining and Knowledge Discovery.
[8] Tsuyoshi Murata,et al. {m , 1934, ACML.
[9] Luciano Floridi,et al. Why a Right to Explanation of Automated Decision-Making Does Not Exist in the General Data Protection Regulation , 2017 .
[10] Brandon M. Greenwell,et al. Interpretable Machine Learning , 2019, Hands-On Machine Learning with R.
[11] Suresh Venkatasubramanian,et al. Auditing black-box models for indirect influence , 2016, Knowledge and Information Systems.
[12] Arvind Satyanarayan,et al. The Building Blocks of Interpretability , 2018 .
[13] G. G. Stokes. "J." , 1890, The New Yale Book of Quotations.
[14] Cynthia Rudin,et al. Supersparse Linear Integer Models for Interpretable Classification , 2013, 1306.6677.
[15] Franco Turini,et al. A Survey of Methods for Explaining Black Box Models , 2018, ACM Comput. Surv..
[16] Zachary Chase Lipton. The mythos of model interpretability , 2016, ACM Queue.
[17] Anas N. Al-Rabadi,et al. A comparison of modified reconstructability analysis and Ashenhurst‐Curtis decomposition of Boolean functions , 2004 .
[18] Yair Zick,et al. Algorithmic Transparency via Quantitative Input Influence: Theory and Experiments with Learning Systems , 2016, 2016 IEEE Symposium on Security and Privacy (SP).
[19] Julia Powles,et al. "Meaningful Information" and the Right to Explanation , 2017, FAT.
[20] Been Kim,et al. Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.
[21] Brad Boehmke,et al. Interpretable Machine Learning , 2019 .
[22] Seth Flaxman,et al. European Union Regulations on Algorithmic Decision-Making and a "Right to Explanation" , 2016, AI Mag..
[23] Daniel G. Goldstein,et al. Manipulating and Measuring Model Interpretability , 2018, CHI.
[24] Leo Breiman,et al. Random Forests , 2001, Machine Learning.
[25] Gilles Louppe,et al. Independent consultant , 2013 .
[26] Cynthia Rudin,et al. Supersparse linear integer models for optimized medical scoring systems , 2015, Machine Learning.
[28] Niklas Lavesson,et al. User-oriented Assessment of Classification Model Understandability , 2011, SCAI.
[29] Emily Chen,et al. How do Humans Understand Explanations from Machine Learning Systems? An Evaluation of the Human-Interpretability of Explanation , 2018, ArXiv.
[30] Mohan S. Kankanhalli,et al. Trends and Trajectories for Explainable, Accountable and Intelligible Systems: An HCI Research Agenda , 2018, CHI.