Tensor2Tensor for Neural Machine Translation
暂无分享,去创建一个
Samy Bengio | Lukasz Kaiser | Ashish Vaswani | Jakob Uszkoreit | Noam Shazeer | François Chollet | Nal Kalchbrenner | Eugene Brevdo | Aidan N. Gomez | Niki Parmar | Llion Jones | Stephan Gouws | Ryan Sepassi | Lukasz Kaiser | Samy Bengio | Ashish Vaswani | Noam M. Shazeer | Nal Kalchbrenner | Llion Jones | E. Brevdo | Niki Parmar | Jakob Uszkoreit | Ryan Sepassi | François Chollet | Stephan Gouws
[1] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[2] Phil Blunsom,et al. Recurrent Continuous Translation Models , 2013, EMNLP.
[3] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[4] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[5] Qun Liu,et al. Encoding Source Language with Convolutional Neural Network for Machine Translation , 2015, ACL.
[6] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[7] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[8] Yuan Yu,et al. TensorFlow: A system for large-scale machine learning , 2016, OSDI.
[9] Samy Bengio,et al. Can Active Memory Replace Attention? , 2016, NIPS.
[10] Alex Graves,et al. Neural Machine Translation in Linear Time , 2016, ArXiv.
[11] Wei Xu,et al. Deep Recurrent Models with Fast-Forward Connections for Neural Machine Translation , 2016, TACL.
[12] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[13] Heiga Zen,et al. WaveNet: A Generative Model for Raw Audio , 2016, SSW.
[14] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[15] Raquel Urtasun,et al. The Reversible Residual Network: Backpropagation Without Storing Activations , 2017, NIPS.
[16] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[17] Geoffrey E. Hinton,et al. Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer , 2017, ICLR.
[18] François Chollet,et al. Xception: Deep Learning with Depthwise Separable Convolutions , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[19] Lukasz Kaiser,et al. Depthwise Separable Convolutions for Neural Machine Translation , 2017, ICLR.
[20] Dustin Tran,et al. Image Transformer , 2018, ICML.