暂无分享,去创建一个
[1] Jamil Zaki,et al. Modeling emotion in complex stories: the Stanford Emotional Narratives Dataset , 2019, ArXiv.
[2] Geoffrey E. Hinton,et al. Grammar as a Foreign Language , 2014, NIPS.
[3] Alexander Binder,et al. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.
[4] Bing Liu,et al. Mining and summarizing customer reviews , 2004, KDD.
[5] Dipanjan Das,et al. BERT Rediscovers the Classical NLP Pipeline , 2019, ACL.
[6] Xiaoli Z. Fern,et al. Interpreting Recurrent and Attention-Based Neural Models: a Case Study on Natural Language Inference , 2018, EMNLP.
[7] Byron C. Wallace,et al. Attention is not Explanation , 2019, NAACL.
[8] Klaus-Robert Müller,et al. Explaining Recurrent Neural Network Predictions in Sentiment Analysis , 2017, WASSA@EMNLP.
[9] Jamil Zaki,et al. Attending to Emotional Narratives , 2019, 2019 8th International Conference on Affective Computing and Intelligent Interaction (ACII).
[10] Christopher D. Manning,et al. Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks , 2015, ACL.
[11] M. Gevrey,et al. Review and comparison of methods to study the contribution of variables in artificial neural network models , 2003 .
[12] Yoav Goldberg,et al. Assessing BERT's Syntactic Abilities , 2019, ArXiv.
[13] Yang Liu,et al. On Identifiability in Transformers , 2020, ICLR.
[14] Willem Zuidema,et al. Quantifying Attention Flow in Transformers , 2020, ACL.
[15] Manaal Faruqui,et al. Attention Interpretability Across NLP Tasks , 2019, ArXiv.
[16] Christopher D. Manning,et al. Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.
[17] Omer Levy,et al. What Does BERT Look at? An Analysis of BERT’s Attention , 2019, BlackboxNLP@ACL.
[18] L. Lin,et al. A concordance correlation coefficient to evaluate reproducibility. , 1989, Biometrics.
[19] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[20] Xinlei Chen,et al. Visualizing and Understanding Neural Models in NLP , 2015, NAACL.
[21] Yiming Yang,et al. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context , 2019, ACL.
[22] Bowen Zhou,et al. A Structured Self-attentive Sentence Embedding , 2017, ICLR.
[23] Li Zhao,et al. Attention-based LSTM for Aspect-level Sentiment Classification , 2016, EMNLP.
[24] Tao Shen,et al. DiSAN: Directional Self-Attention Network for RNN/CNN-free Language Understanding , 2017, AAAI.
[25] Rico Sennrich,et al. Context-Aware Neural Machine Translation Learns Anaphora Resolution , 2018, ACL.
[26] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[27] Jörg Tiedemann,et al. An Analysis of Encoder Representations in Transformer-Based Machine Translation , 2018, BlackboxNLP@EMNLP.
[28] Christopher Potts,et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.
[29] Yannis Dimopoulos,et al. Use of some sensitivity criteria for choosing networks with good generalization ability , 1995, Neural Processing Letters.
[30] Andrew Zisserman,et al. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.
[31] Amy Beth Warriner,et al. Norms of valence, arousal, and dominance for 13,915 English lemmas , 2013, Behavior Research Methods.
[32] Yonatan Belinkov,et al. Analyzing the Structure of Attention in a Transformer Language Model , 2019, BlackboxNLP@ACL.
[33] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[34] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[35] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[36] Fedor Moiseev,et al. Analyzing Multi-Head Self-Attention: Specialized Heads Do the Heavy Lifting, the Rest Can Be Pruned , 2019, ACL.
[37] Noah A. Smith,et al. Is Attention Interpretable? , 2019, ACL.
[38] Yang Liu,et al. Learning Structured Text Representations , 2017, TACL.
[39] Fred Popowich,et al. QA2Explanation: Generating and Evaluating Explanations for Question Answering Systems over Knowledge Graph , 2018, INTEXSEMPAR.
[40] Christopher D. Manning,et al. A Structural Probe for Finding Syntax in Word Representations , 2019, NAACL.