Insertion Transformer: Flexible Sequence Generation via Insertion Operations
暂无分享,去创建一个
Jakob Uszkoreit | William Chan | Jamie Kiros | Mitchell Stern | Jakob Uszkoreit | William Chan | J. Kiros | Mitchell Stern
[1] Alex Graves,et al. Sequence Transduction with Recurrent Neural Networks , 2012, ArXiv.
[2] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[3] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[4] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[5] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[6] Christopher D. Manning,et al. Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.
[7] Samy Bengio,et al. Show and tell: A neural image caption generator , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[8] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[9] Alex Graves,et al. Conditional Image Generation with PixelCNN Decoders , 2016, NIPS.
[10] Razvan Pascanu,et al. Policy Distillation , 2015, ICLR.
[11] Dale Schuurmans,et al. Reward Augmented Maximum Likelihood for Neural Structured Prediction , 2016, NIPS.
[12] Quoc V. Le,et al. Listen, attend and spell: A neural network for large vocabulary conversational speech recognition , 2015, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[13] Alexander M. Rush,et al. Sequence-Level Knowledge Distillation , 2016, EMNLP.
[14] Yoshua Bengio,et al. End-to-end attention-based large vocabulary speech recognition , 2015, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[15] Samy Bengio,et al. Order Matters: Sequence to sequence for sets , 2015, ICLR.
[16] Heiga Zen,et al. WaveNet: A Generative Model for Raw Audio , 2016, SSW.
[17] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[18] Sergio Gomez Colmenarejo,et al. Parallel Multiscale Autoregressive Density Estimation , 2017, ICML.
[19] Samy Bengio,et al. Tacotron: Towards End-to-End Speech Synthesis , 2017, INTERSPEECH.
[20] Victor O. K. Li,et al. Non-Autoregressive Neural Machine Translation , 2017, ICLR.
[21] Jakob Uszkoreit,et al. Blockwise Parallel Decoding for Deep Autoregressive Models , 2018, NeurIPS.
[22] Samy Bengio,et al. Tensor2Tensor for Neural Machine Translation , 2018, AMTA.
[23] Ji Zhang,et al. Semi-Autoregressive Neural Machine Translation , 2018, EMNLP.
[24] Jason Lee,et al. Deterministic Non-Autoregressive Neural Sequence Modeling by Iterative Refinement , 2018, EMNLP.
[25] David Barber,et al. Generating Sentences Using a Dynamic Canvas , 2018, AAAI.
[26] Mohammad Norouzi,et al. The Importance of Generation Order in Language Modeling , 2018, EMNLP.
[27] Ruslan Salakhutdinov,et al. Breaking the Softmax Bottleneck: A High-Rank RNN Language Model , 2017, ICLR.
[28] Qi Liu,et al. Insertion-based Decoding with Automatically Inferred Generation Order , 2019, Transactions of the Association for Computational Linguistics.
[29] Kyunghyun Cho,et al. Appendix : Non-Monotonic Sequential Text Generation , 2019 .