暂无分享,去创建一个
Shijie Geng | Shuchang Liu | Yingqiang Ge | Yongfeng Zhang | Fei Sun | Zelong Li | Yunqi Li | Shuyuan Xu | Juntao Tan
[1] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[2] Byron C. Wallace,et al. ERASER: A Benchmark to Evaluate Rationalized NLP Models , 2020, ACL.
[3] Mohit Bansal,et al. Evaluating Explainable AI: Which Algorithmic Explanations Help Users Predict Model Behavior? , 2020, ACL.
[4] Andrew Zisserman,et al. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.
[5] Been Kim,et al. Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.
[6] Danqi Chen,et al. of the Association for Computational Linguistics: , 2001 .
[7] George Chrysostomou,et al. Variable Instance-Level Explainability for Text Classification , 2021, ArXiv.
[8] William W. Cohen,et al. Evaluating Explanations: How Much Do Explanations from the Teacher Aid Students? , 2020, TACL.
[9] LiuHuan,et al. Causal Interpretability for Machine Learning - Problems, Methods and Evaluation , 2020 .
[10] Tommi S. Jaakkola,et al. Rethinking Cooperative Rationalization: Introspective Extraction and Complement Control , 2019, EMNLP.
[11] Carlos Guestrin,et al. Anchors: High-Precision Model-Agnostic Explanations , 2018, AAAI.
[12] M. Kendall. The treatment of ties in ranking problems. , 1945, Biometrika.
[13] Pietro Liò,et al. Graph Attention Networks , 2017, ICLR.
[14] Cengiz Öztireli,et al. Towards better understanding of gradient-based attribution methods for Deep Neural Networks , 2017, ICLR.
[15] Jasmijn Bastings,et al. The elephant in the interpretability room: Why use attention as explanation when we have saliency methods? , 2020, BLACKBOXNLP.
[16] Ziyan Wu,et al. Counterfactual Visual Explanations , 2019, ICML.
[17] Yuval Pinter,et al. Attention is not not Explanation , 2019, EMNLP.
[18] Percy Liang,et al. Understanding Black-box Predictions via Influence Functions , 2017, ICML.
[19] Chris Russell,et al. Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR , 2017, ArXiv.
[20] Martin Wattenberg,et al. Interpretability Beyond Feature Attribution: Quantitative Testing with Concept Activation Vectors (TCAV) , 2017, ICML.
[21] Philipp Koehn,et al. Evaluating Saliency Methods for Neural Language Models , 2021, NAACL.
[22] Tommi S. Jaakkola,et al. A causal framework for explaining the predictions of black-box sequence-to-sequence models , 2017, EMNLP.
[23] Amit Sharma,et al. Explaining machine learning classifiers through diverse counterfactual explanations , 2020, FAT*.
[24] Xinlei Chen,et al. Visualizing and Understanding Neural Models in NLP , 2015, NAACL.
[25] Dong Nguyen,et al. Comparing Automatic and Human Evaluation of Local Explanations for Text Classification , 2018, NAACL.
[26] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[27] Bernease Herman,et al. The Promise and Peril of Human Evaluation for Model Interpretability , 2017, ArXiv.
[28] John P. Dickerson,et al. Counterfactual Explanations for Machine Learning: A Review , 2020, ArXiv.
[29] Bo Pang,et al. Thumbs up? Sentiment Classification using Machine Learning Techniques , 2002, EMNLP.
[30] Christine D. Piatko,et al. Using “Annotator Rationales” to Improve Machine Learning for Text Categorization , 2007, NAACL.
[31] Greg Durrett,et al. Evaluating Explanations for Reading Comprehension with Realistic Counterfactuals , 2021, ArXiv.
[32] Marko Robnik-Sikonja,et al. Explaining Classifications For Individual Instances , 2008, IEEE Transactions on Knowledge and Data Engineering.
[33] Yoav Goldberg,et al. Towards Faithfully Interpretable NLP Systems: How Should We Define and Evaluate Faithfulness? , 2020, ACL.