2 Background : Latent Alignment and Neural Attention
暂无分享,去创建一个
[1] Sunita Sarawagi,et al. Surprisingly Easy Hard-Attention for Sequence to Sequence Learning , 2018, EMNLP.
[2] Ryan Cotterell,et al. Hard Non-Monotonic Attention for Character-Level Transduction , 2018, EMNLP.
[3] Martin Jankowiak,et al. Pathwise Derivatives Beyond the Reparameterization Trick , 2018, ICML.
[4] Wilker Aziz,et al. A Stochastic Decoder for Neural Machine Translation , 2018, ACL.
[5] Jason Lee,et al. Deterministic Non-Autoregressive Neural Sequence Modeling by Iterative Refinement , 2018, EMNLP.
[6] Claire Cardie,et al. SparseMAP: Differentiable Sparse Structured Inference , 2018, ICML.
[7] Arthur Mensch,et al. Differentiable Dynamic Programming for Structured Prediction and Attention , 2018, ICML.
[8] Alexander M. Rush,et al. Semi-Amortized Variational Autoencoders , 2018, ICML.
[9] Shan Wu,et al. Variational Recurrent Neural Machine Translation , 2018, AAAI.
[10] Pascal Poupart,et al. Variational Attention for Sequence-to-Sequence Models , 2017, COLING.
[11] Marc'Aurelio Ranzato,et al. Classical Structured Prediction Losses for Sequence to Sequence Learning , 2017, NAACL.
[12] David Duvenaud,et al. Backpropagation through the Void: Optimizing control variates for black-box gradient estimation , 2017, ICLR.
[13] Matthew D. Hoffman,et al. On the challenges of learning with inference networks on sparse, high-dimensional data , 2017, AISTATS.
[14] Lei Zhang,et al. Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[15] Chong Wang,et al. Towards Neural Phrase-based Machine Translation , 2017, ICLR.
[16] Yang Liu,et al. Learning Structured Text Representations , 2017, TACL.
[17] Colin Raffel,et al. Learning Hard Alignments with Variational Inference , 2017, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[18] Yoshua Bengio,et al. Z-Forcing: Training Stochastic Recurrent Networks , 2017, NIPS.
[19] Jörg Bornschein,et al. Variational Memory Addressing in Generative Models , 2017, NIPS.
[20] Kewei Tu,et al. Structured Attentions for Visual Question Answering , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[21] Tommi S. Jaakkola,et al. A causal framework for explaining the predictions of black-box sequence-to-sequence models , 2017, EMNLP.
[22] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[23] Bonggun Shin,et al. Classification of radiology reports using neural attention models , 2017, 2017 International Joint Conference on Neural Networks (IJCNN).
[24] Vlad Niculae,et al. A Regularized Framework for Sparse and Structured Neural Attention , 2017, NIPS.
[25] Jascha Sohl-Dickstein,et al. REBAR: Low-variance, unbiased gradient estimates for discrete latent variable models , 2017, NIPS.
[26] Charles A. Sutton,et al. Autoencoding Variational Inference For Topic Models , 2017, ICLR.
[27] Alexander M. Rush,et al. Structured Attention Networks , 2017, ICLR.
[28] Lei Yu,et al. The Neural Noisy Channel , 2016, ICLR.
[29] Ben Poole,et al. Categorical Reparameterization with Gumbel-Softmax , 2016, ICLR.
[30] Yee Whye Teh,et al. The Concrete Distribution: A Continuous Relaxation of Discrete Random Variables , 2016, ICLR.
[31] Uri Shalit,et al. Structured Inference Networks for Nonlinear State Space Models , 2016, AAAI.
[32] Yaoliang Yu,et al. Dropout with Expectation-linear Regularization , 2016, ICLR.
[33] Alexander M. Rush,et al. Image-to-Markup Generation with Coarse-to-Fine Attention , 2016, ICML.
[34] Joelle Pineau,et al. An Actor-Critic Algorithm for Sequence Prediction , 2016, ICLR.
[35] Joelle Pineau,et al. A Hierarchical Latent Variable Encoder-Decoder Model for Generating Dialogues , 2016, AAAI.
[36] A. Algorithms. Online and Linear-Time Attention by Enforcing Monotonic Alignments , 2017 .
[37] Roman Novak,et al. Iterative Refinement for Machine Translation , 2016, ArXiv.
[38] Lei Yu,et al. Online Segment to Segment Neural Transduction , 2016, EMNLP.
[39] Yoshua Bengio,et al. Dynamic Neural Turing Machine with Soft and Hard Addressing Schemes , 2016, ArXiv.
[40] Regina Barzilay,et al. Rationalizing Neural Predictions , 2016, EMNLP.
[41] Alexander M. Rush,et al. Sequence-to-Sequence Learning as Beam-Search Optimization , 2016, EMNLP.
[42] Min Zhang,et al. Variational Neural Machine Translation , 2016, EMNLP.
[43] Ole Winther,et al. Sequential Neural Models with Stochastic Layers , 2016, NIPS.
[44] Hang Li,et al. “ Tony ” DNN Embedding for “ Tony ” Selective Read for “ Tony ” ( a ) Attention-based Encoder-Decoder ( RNNSearch ) ( c ) State Update s 4 SourceVocabulary Softmax Prob , 2016 .
[45] Andriy Mnih,et al. Variational Inference for Monte Carlo Objectives , 2016, ICML.
[46] Ramón Fernández Astudillo,et al. From Softmax to Sparsemax: A Sparse Model of Attention and Multi-Label Classification , 2016, ICML.
[47] Yang Liu,et al. Modeling Coverage for Neural Machine Translation , 2016, ACL.
[48] Gholamreza Haffari,et al. Incorporating Structural Alignment Biases into an Attentional Neural Translation Model , 2016, NAACL.
[49] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[50] Alexander J. Smola,et al. Stacked Attention Networks for Image Question Answering , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[51] Phil Blunsom,et al. Reasoning about Entailment with Neural Attention , 2015, ICLR.
[52] Ruslan Salakhutdinov,et al. Importance Weighted Autoencoders , 2015, ICLR.
[53] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[54] Brendan J. Frey,et al. Learning Wake-Sleep Recurrent Attention Models , 2015, NIPS.
[55] Jason Weston,et al. A Neural Attention Model for Abstractive Sentence Summarization , 2015, EMNLP.
[56] Christopher D. Manning,et al. Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.
[57] Quoc V. Le,et al. Listen, Attend and Spell , 2015, ArXiv.
[58] Yoshua Bengio,et al. Describing Multimedia Content Using Attention-Based Encoder-Decoder Networks , 2015, IEEE Transactions on Multimedia.
[59] Yoshua Bengio,et al. Attention-Based Models for Speech Recognition , 2015, NIPS.
[60] Yoshua Bengio,et al. A Recurrent Latent Variable Model for Sequential Data , 2015, NIPS.
[61] Kaiming He,et al. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[62] Jason Weston,et al. End-To-End Memory Networks , 2015, NIPS.
[63] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[64] Koray Kavukcuoglu,et al. Multiple Object Recognition with Visual Attention , 2014, ICLR.
[65] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[66] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[67] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[68] Alex Graves,et al. Recurrent Models of Visual Attention , 2014, NIPS.
[69] Karol Gregor,et al. Neural Variational Inference and Learning in Belief Networks , 2014, ICML.
[70] Daan Wierstra,et al. Stochastic Backpropagation and Approximate Inference in Deep Generative Models , 2014, ICML.
[71] Marcello Federico,et al. Report on the 11th IWSLT evaluation campaign , 2014, IWSLT.
[72] Noah A. Smith,et al. A Simple, Fast, and Effective Reparameterization of IBM Model 2 , 2013, NAACL.
[73] Chong Wang,et al. Stochastic variational inference , 2012, J. Mach. Learn. Res..
[74] Philipp Koehn,et al. Moses: Open Source Toolkit for Statistical Machine Translation , 2007, ACL.
[75] R. J. Williams,et al. Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.
[76] Robert L. Mercer,et al. The Mathematics of Statistical Machine Translation: Parameter Estimation , 1993, CL.