Why is Attention Not So Interpretable
暂无分享,去创建一个
Fei Wang | Hao Li | Jian Liang | Bing Bai | Guanhua Zhang | Kun Bai
[1] Jimeng Sun,et al. RETAIN: An Interpretable Predictive Model for Healthcare using Reverse Time Attention Mechanism , 2016, NIPS.
[2] Noah A. Smith,et al. Is Attention Interpretable? , 2019, ACL.
[3] Hao Li,et al. CSRN: Collaborative Sequential Recommendation Networks for News Retrieval , 2020, ArXiv.
[4] Zachary C. Lipton,et al. The mythos of model interpretability , 2018, Commun. ACM.
[5] D. Rubin,et al. The central role of the propensity score in observational studies for causal effects , 1983 .
[6] Xia Hu,et al. Techniques for interpretable machine learning , 2018, Commun. ACM.
[7] Guan Huang,et al. Attention-Guided Unified Network for Panoptic Segmentation , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[8] Le Song,et al. Learning to Explain: An Information-Theoretic Perspective on Model Interpretation , 2018, ICML.
[9] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[10] Walter Karlen,et al. CXPlain: Causal Explanations for Model Interpretation under Uncertainty , 2019, NeurIPS.
[11] Wei Wu,et al. Explaining a black-box using Deep Variational Information Bottleneck Approach , 2019, ArXiv.
[12] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[13] Nenghai Yu,et al. Learning Spatial Regularization with Image-Level Supervisions for Multi-label Image Classification , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[14] Siu Cheung Hui,et al. Multi-Pointer Co-Attention Networks for Recommendation , 2018, KDD.
[15] Jun Zhao,et al. Recurrent Convolutional Neural Networks for Text Classification , 2015, AAAI.
[16] Ben Poole,et al. Categorical Reparametrization with Gumble-Softmax , 2017, ICLR 2017.
[17] Xiang Zhang,et al. Character-level Convolutional Networks for Text Classification , 2015, NIPS.
[18] Tiejun Zhao,et al. Selection Bias Explorations and Debias Methods for Natural Language Sentence Matching Datasets , 2019, ACL.
[19] Christopher Winship,et al. THE ESTIMATION OF CAUSAL EFFECTS FROM OBSERVATIONAL DATA , 1999 .
[20] Bing Bai,et al. Adversarial Infidelity Learning for Model Interpretation , 2020, KDD.
[21] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[22] Vladimir Vapnik,et al. Principles of Risk Minimization for Learning Theory , 1991, NIPS.
[23] Byron C. Wallace,et al. Attention is not Explanation , 2019, NAACL.
[24] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[25] Andrew Zisserman,et al. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.
[26] Fei Sha,et al. Aligning Where to See and What to Tell: Image Captioning with Region-Based Attention and Scene-Specific Contexts , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[27] Tommi S. Jaakkola,et al. Rethinking Cooperative Rationalization: Introspective Extraction and Complement Control , 2019, EMNLP.
[28] Ankur Taly,et al. Axiomatic Attribution for Deep Networks , 2017, ICML.
[29] Ramón Fernández Astudillo,et al. From Softmax to Sparsemax: A Sparse Model of Attention and Multi-Label Classification , 2016, ICML.
[30] Thorsten Joachims,et al. Unbiased Learning-to-Rank with Biased Feedback , 2016, WSDM.
[31] Fei Wang,et al. General-Purpose User Embeddings based on Mobile App Usage , 2020, KDD.
[32] Omer Levy,et al. What Does BERT Look at? An Analysis of BERT’s Attention , 2019, BlackboxNLP@ACL.
[33] Peter M. Steiner,et al. Can Nonrandomized Experiments Yield Accurate Answers? A Randomized Experiment Comparing Random and Nonrandom Assignments , 2008 .
[34] Christopher Potts,et al. Learning Word Vectors for Sentiment Analysis , 2011, ACL.
[35] Conghui Zhu,et al. Demographics Should Not Be the Reason of Toxicity: Mitigating Discrimination in Text Classifications with Instance Weighting , 2020, ACL.
[36] D. Rubin. Estimating causal effects of treatments in randomized and nonrandomized studies. , 1974 .
[37] Ashkan Ertefaie,et al. Comparing Approaches to Causal Inference for Longitudinal Data: Inverse Probability Weighting versus Propensity Scores , 2010, The international journal of biostatistics.
[38] Dumitru Erhan,et al. A Benchmark for Interpretability Methods in Deep Neural Networks , 2018, NeurIPS.
[39] Philip S. Yu,et al. An improved categorization of classifier's sensitivity on sample selection bias , 2005, Fifth IEEE International Conference on Data Mining (ICDM'05).
[40] Bianca Zadrozny,et al. Learning and evaluating classifiers under sample selection bias , 2004, ICML.
[41] Fei Wang,et al. Should Health Care Demand Interpretable Artificial Intelligence or Accept “Black Box” Medicine? , 2019, Annals of Internal Medicine.
[42] Yuval Pinter,et al. Attention is not not Explanation , 2019, EMNLP.
[43] Geoffrey E. Hinton,et al. Grammar as a Foreign Language , 2014, NIPS.
[44] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[45] Li Zhao,et al. Attention-based LSTM for Aspect-level Sentiment Classification , 2016, EMNLP.
[46] Denali Molitor,et al. Model Agnostic Supervised Local Explanations , 2018, NeurIPS.
[47] Subbarao Kambhampati,et al. Explicability? Legibility? Predictability? Transparency? Privacy? Security? The Emerging Landscape of Interpretable Agent Behavior , 2018, ICAPS.
[48] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[49] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.