暂无分享,去创建一个
Krishnaram Kenthapadi | Muhammad Bilal Zafar | Michele Donini | C'edric Archambeau | Dylan Slack | Sanjiv Das
[1] Suresh Venkatasubramanian,et al. Problems with Shapley-value-based explanations as feature importance measures , 2020, ICML.
[2] Himabindu Lakkaraju,et al. Robust and Stable Black Box Explanations , 2020, ICML.
[3] Franco Turini,et al. A Survey of Methods for Explaining Black Box Models , 2018, ACM Comput. Surv..
[4] Philipp Koehn,et al. Saliency-driven Word Alignment Interpretation for Neural Machine Translation , 2019, WMT.
[5] Byron C. Wallace,et al. ERASER: A Benchmark to Evaluate Rationalized NLP Models , 2020, ACL.
[6] Been Kim,et al. Sanity Checks for Saliency Maps , 2018, NeurIPS.
[7] Sameer Singh,et al. Fooling LIME and SHAP: Adversarial Attacks on Post hoc Explanation Methods , 2020, AIES.
[8] Pekka Korhonen,et al. Good debt or bad debt: Detecting semantic orientations in economic texts , 2013, J. Assoc. Inf. Sci. Technol..
[9] Alexander Binder,et al. Evaluating the Visualization of What a Deep Neural Network Has Learned , 2015, IEEE Transactions on Neural Networks and Learning Systems.
[10] Christopher Potts,et al. Learning Word Vectors for Sentiment Analysis , 2011, ACL.
[11] Alexandra Chouldechova,et al. Bias in Bios: A Case Study of Semantic Representation Bias in a High-Stakes Setting , 2019, FAT.
[12] Felix Bießmann,et al. Quantifying Interpretability and Trust in Machine Learning Systems , 2019, ArXiv.
[13] Xuanjing Huang,et al. How to Fine-Tune BERT for Text Classification? , 2019, CCL.
[14] Klaus-Robert Müller,et al. Fairwashing Explanations with Off-Manifold Detergent , 2020, ICML.
[15] Martin Wattenberg,et al. SmoothGrad: removing noise by adding noise , 2017, ArXiv.
[16] Shi Feng,et al. Pathologies of Neural Models Make Interpretations Difficult , 2018, EMNLP.
[17] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[18] Prudhvi Gurram,et al. Sanity Checks for Saliency Metrics , 2019, AAAI.
[19] Andrew Zisserman,et al. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.
[20] Albert Gordo,et al. Learning Global Additive Explanations for Neural Nets Using Model Distillation , 2018 .
[21] Ankur Taly,et al. Explainable machine learning in deployment , 2020, FAT*.
[22] Andrea Vedaldi,et al. Understanding Deep Networks via Extremal Perturbations and Smooth Masks , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[23] Gaberell Drachman,et al. On explaining “Explanations” , 2016 .
[24] Sameer Singh,et al. AllenNLP Interpret: A Framework for Explaining Predictions of NLP Models , 2019, EMNLP.
[25] Chong Wang,et al. Reading Tea Leaves: How Humans Interpret Topic Models , 2009, NIPS.
[26] Been Kim,et al. Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.
[27] Lysandre Debut,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[28] Jakob Grue Simonsen,et al. A Diagnostic Study of Explainability Techniques for Text Classification , 2020, EMNLP.
[29] Somesh Jha,et al. Concise Explanations of Neural Networks using Adversarial Training , 2018, ICML.
[30] Avanti Shrikumar,et al. Learning Important Features Through Propagating Activation Differences , 2017, ICML.
[31] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[32] Ilaria Liccardi,et al. Debugging Tests for Model Explanations , 2020, NeurIPS.
[33] Klaus-Robert Müller,et al. Explaining Predictions of Non-Linear Classifiers in NLP , 2016, Rep4NLP@ACL.
[34] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[35] Tommi S. Jaakkola,et al. Towards Robust Interpretability with Self-Explaining Neural Networks , 2018, NeurIPS.
[36] Xinlei Chen,et al. Visualizing and Understanding Neural Models in NLP , 2015, NAACL.
[37] Somesh Jha,et al. Robust Attribution Regularization , 2019, NeurIPS.
[38] Dong Nguyen,et al. Comparing Automatic and Human Evaluation of Local Explanations for Text Classification , 2018, NAACL.
[39] Scott Lundberg,et al. A Unified Approach to Interpreting Model Predictions , 2017, NIPS.
[40] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[41] Frank Hutter,et al. Decoupled Weight Decay Regularization , 2017, ICLR.
[42] Ankur Taly,et al. Axiomatic Attribution for Deep Networks , 2017, ICML.
[43] Klaus-Robert Müller,et al. Explanations can be manipulated and geometry is to blame , 2019, NeurIPS.
[44] Abubakar Abid,et al. Interpretation of Neural Networks is Fragile , 2017, AAAI.
[45] Cuntai Guan,et al. A Survey on Explainable Artificial Intelligence (XAI): Toward Medical XAI , 2019, IEEE Transactions on Neural Networks and Learning Systems.
[46] Daniel G. Goldstein,et al. Manipulating and Measuring Model Interpretability , 2018, CHI.
[47] Christopher Potts,et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.
[48] Martin Wattenberg,et al. Interpretability Beyond Feature Attribution: Quantitative Testing with Concept Activation Vectors (TCAV) , 2017, ICML.
[49] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[50] Mohit Bansal,et al. Evaluating Explainable AI: Which Algorithmic Explanations Help Users Predict Model Behavior? , 2020, ACL.
[51] Mukund Sundararajan,et al. The many Shapley values for model explanation , 2019, ICML.
[52] Lalana Kagal,et al. Explaining Explanations: An Overview of Interpretability of Machine Learning , 2018, 2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA).