Deliberation Networks: Sequence Generation Beyond One-Pass Decoding
暂无分享,去创建一个
Nenghai Yu | Lijun Wu | Tao Qin | Tie-Yan Liu | Fei Tian | Jianxin Lin | Yingce Xia | Tie-Yan Liu | Tao Qin | Fei Tian | Lijun Wu | Yingce Xia | Nenghai Yu | Jianxin Lin
[1] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[2] Marc'Aurelio Ranzato,et al. Sequence Level Training with Recurrent Neural Networks , 2015, ICLR.
[3] Samy Bengio,et al. Show and tell: A neural image caption generator , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[4] Li Zhao,et al. Sequence Prediction with Unlabeled Data by Reward Function Learning , 2017, IJCAI.
[5] Nenghai Yu,et al. Dual Inference for Machine Learning , 2017, IJCAI.
[6] Matthew D. Zeiler. ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.
[7] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.
[8] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[9] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[10] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[11] Di He,et al. Decoding with Value Networks for Neural Machine Translation , 2017, NIPS.
[12] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[13] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[14] Nenghai Yu,et al. Dual Supervised Learning , 2017, ICML.
[15] Yang Liu,et al. Minimum Risk Training for Neural Machine Translation , 2015, ACL.
[16] Yang Liu,et al. Neural Machine Translation with Reconstruction , 2016, AAAI.
[17] Alexander M. Rush,et al. Sequence-to-Sequence Learning as Beam-Search Optimization , 2016, EMNLP.
[18] Yoshua Bengio,et al. On Using Very Large Target Vocabulary for Neural Machine Translation , 2014, ACL.
[19] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[20] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[21] Jason Weston,et al. A Neural Attention Model for Abstractive Sentence Summarization , 2015, EMNLP.
[22] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[23] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[24] John Salvatier,et al. Theano: A Python framework for fast computation of mathematical expressions , 2016, ArXiv.
[25] Ye Yuan,et al. Review Networks for Caption Generation , 2016, NIPS.
[26] Nenghai Yu,et al. Sequence Generation with Target Attention , 2017, ECML/PKDD.
[27] Daniel Jurafsky,et al. A Simple, Fast Diverse Decoding Algorithm for Neural Generation , 2016, ArXiv.
[28] Tie-Yan Liu,et al. Dual Learning for Machine Translation , 2016, NIPS.
[29] Matteo Negri,et al. The FBK Participation in the WMT 2016 Automatic Post-editing Shared Task , 2015, WMT.
[30] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[31] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[32] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[33] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[34] Jan Niehues,et al. Pre-Translation for Neural Machine Translation , 2016, COLING.