暂无分享,去创建一个
[1] Catherine Blake,et al. UCI Repository of machine learning databases , 1998 .
[2] Wojciech Zaremba,et al. Improved Techniques for Training GANs , 2016, NIPS.
[3] Klaus Broelemann,et al. Learning Model-Agnostic Counterfactual Explanations for Tabular Data , 2019, WWW.
[4] Johannes Gehrke,et al. Intelligible Models for HealthCare: Predicting Pneumonia Risk and Hospital 30-day Readmission , 2015, KDD.
[5] Martin Hlosta,et al. Open University Learning Analytics dataset , 2017, Scientific Data.
[6] Jette Henderson,et al. CERTIFAI: A Common Framework to Provide Explanations and Analyse the Fairness and Robustness of Black-box Models , 2020, AIES.
[7] Amit Sharma,et al. Explaining machine learning classifiers through diverse counterfactual explanations , 2020, FAT*.
[8] Tianqi Chen,et al. Empirical Evaluation of Rectified Activations in Convolutional Network , 2015, ArXiv.
[9] Yang Liu,et al. Actionable Recourse in Linear Classification , 2018, FAT.
[10] John P. Dickerson,et al. Counterfactual Explanations for Machine Learning: A Review , 2020, ArXiv.
[11] Jun Zhao,et al. 'It's Reducing a Human Being to a Percentage': Perceptions of Justice in Algorithmic Decisions , 2018, CHI.
[12] Cynthia Rudin,et al. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead , 2018, Nature Machine Intelligence.
[13] Scott Lundberg,et al. A Unified Approach to Interpreting Model Predictions , 2017, NIPS.
[14] Amit Sharma,et al. Preserving Causal Constraints in Counterfactual Explanations for Machine Learning Classifiers , 2019, ArXiv.
[15] Jure Leskovec,et al. Interpretable Decision Sets: A Joint Framework for Description and Prediction , 2016, KDD.
[16] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[17] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[18] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[19] Amit Dhurandhar,et al. Explanations based on the Missing: Towards Contrastive Explanations with Pertinent Negatives , 2018, NeurIPS.
[20] Ankur Taly,et al. Explainable machine learning in deployment , 2020, FAT*.
[21] Janis Klaise,et al. Interpretable Counterfactual Explanations Guided by Prototypes , 2019, ECML/PKDD.
[22] Cynthia Rudin,et al. This Looks Like That: Deep Learning for Interpretable Image Recognition , 2018 .
[23] Paulo Cortez,et al. Using data mining to predict secondary school student performance , 2008 .
[24] Bernhard Schölkopf,et al. Algorithmic Recourse: from Counterfactual Explanations to Interventions , 2020, FAccT.
[25] Trade-Offs between Fairness and Interpretability in Machine Learning , 2020 .
[26] Johannes Gehrke,et al. Accurate intelligible models with pairwise interactions , 2013, KDD.
[27] Chris Russell,et al. Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR , 2017, ArXiv.