Saliency-driven Word Alignment Interpretation for Neural Machine Translation
暂无分享,去创建一个
[1] Paul Buitelaar,et al. Enhancing statistical machine translation with bilingual terminology in a CAT environment , 2014, AMTA.
[2] Andrew Zisserman,et al. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.
[3] Philipp Koehn,et al. Six Challenges for Neural Machine Translation , 2017, NMT@ACL.
[4] Walter Schneider,et al. Controlled and automatic human information processing: II. Perceptual learning, automatic attending and a general theory. , 1977 .
[5] Daniel Marcu,et al. Statistical Phrase-Based Translation , 2003, NAACL.
[6] R. Shiffrin,et al. Controlled and automatic human information processing: I , 1977 .
[7] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[8] Yoshua Bengio,et al. Attention-Based Models for Speech Recognition , 2015, NIPS.
[9] Thomas Brox,et al. Striving for Simplicity: The All Convolutional Net , 2014, ICLR.
[10] Wojciech Samek,et al. Methods for interpreting and understanding deep neural networks , 2017, Digit. Signal Process..
[11] Jason Weston,et al. A Neural Attention Model for Abstractive Sentence Summarization , 2015, EMNLP.
[12] Jakob Uszkoreit,et al. A Decomposable Attention Model for Natural Language Inference , 2016, EMNLP.
[13] Christof Monz,et al. What does Attention in Neural Machine Translation Pay Attention to? , 2017, IJCNLP.
[14] Christopher D. Manning,et al. Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.
[15] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[16] Noah A. Smith,et al. A Simple, Fast, and Effective Reparameterization of IBM Model 2 , 2013, NAACL.
[17] Gonzalo Iglesias,et al. Neural Machine Translation Decoding with Terminology Constraints , 2018, NAACL.
[18] Lemao Liu,et al. Neural Machine Translation with Supervised Attention , 2016, COLING.
[19] Hermann Ney,et al. A Systematic Comparison of Various Statistical Alignment Models , 2003, CL.
[20] Robert L. Mercer,et al. The Mathematics of Statistical Machine Translation: Parameter Estimation , 1993, CL.
[21] Shujian Huang,et al. Improved Neural Machine Translation with a Syntax-Aware Encoder and Decoder , 2017, ACL.
[22] Christopher D. Manning,et al. Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.
[23] Ronan Collobert,et al. Neural Network-based Word Alignment through Score Aggregation , 2016, WMT.
[24] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[25] Hermann Ney,et al. Alignment-Based Neural Machine Translation , 2016, WMT.
[26] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[27] Xinlei Chen,et al. Visualizing and Understanding Neural Models in NLP , 2015, NAACL.
[28] Jörg Tiedemann,et al. An Analysis of Encoder Representations in Transformer-Based Machine Translation , 2018, BlackboxNLP@EMNLP.
[29] Ankur Taly,et al. Axiomatic Attribution for Deep Networks , 2017, ICML.
[30] Hermann Ney,et al. On The Alignment Problem In Multi-Head Attention-Based Neural Machine Translation , 2018, WMT.
[31] Joakim Nivre,et al. An Analysis of Attention Mechanisms: The Case of Word Sense Disambiguation in Neural Machine Translation , 2018, WMT.
[32] Martin Wattenberg,et al. SmoothGrad: removing noise by adding noise , 2017, ArXiv.
[33] Quoc V. Le,et al. Listen, attend and spell: A neural network for large vocabulary conversational speech recognition , 2015, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[34] Yang Liu,et al. Visualizing and Understanding Neural Machine Translation , 2017, ACL.
[35] Rico Sennrich,et al. Why Self-Attention? A Targeted Evaluation of Neural Machine Translation Architectures , 2018, EMNLP.
[36] John DeNero,et al. Adding Interpretable Attention to Neural Translation Models Improves Word Alignment , 2019, ArXiv.
[37] Zhiguo Wang,et al. Supervised Attentions for Neural Machine Translation , 2016, EMNLP.
[38] Xiaoli Z. Fern,et al. Interpreting Recurrent and Attention-Based Neural Models: a Case Study on Natural Language Inference , 2018, EMNLP.
[39] Hermann Ney,et al. Neural Hidden Markov Model for Machine Translation , 2018, ACL.
[40] Alexander Binder,et al. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.
[41] David Chiang,et al. Improving Lexical Choice in Neural Machine Translation , 2017, NAACL.
[42] Jun-Seok Kim,et al. Interactive Visualization and Manipulation of Attention-based Neural Machine Translation , 2017, EMNLP.