Machine Translation Decoding beyond Beam Search
暂无分享,去创建一个
Oriol Vinyals | Karen Simonyan | Ioannis Antonoglou | Laurent Sifre | Jean-Baptiste Alayrac | Jean-Baptiste Lespiau | Miruna Pislar | R'emi Leblond | Oriol Vinyals | L. Sifre | Rémi Leblond | Ioannis Antonoglou | K. Simonyan | J. Lespiau | Jean-Baptiste Alayrac | Miruna Pislar
[1] Gabriel Synnaeve,et al. A Fully Differentiable Beam Search Decoder , 2019, ICML.
[2] J. Christopher Beck,et al. Empirical Analysis of Beam Search Performance Degradation in Neural Sequence Models , 2019, ICML.
[3] Yann Dauphin,et al. Hierarchical Neural Story Generation , 2018, ACL.
[4] Alon Lavie,et al. Meteor Universal: Language Specific Translation Evaluation for Any Target Language , 2014, WMT@ACL.
[5] Kilian Q. Weinberger,et al. BERTScore: Evaluating Text Generation with BERT , 2019, ICLR.
[6] Joelle Pineau,et al. An Actor-Critic Algorithm for Sequence Prediction , 2016, ICLR.
[7] Di He,et al. Decoding with Value Networks for Neural Machine Translation , 2017, NIPS.
[8] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[9] Csaba Szepesvári,et al. Bandit Based Monte-Carlo Planning , 2006, ECML.
[10] Simon M. Lucas,et al. A Survey of Monte Carlo Tree Search Methods , 2012, IEEE Transactions on Computational Intelligence and AI in Games.
[11] Ramón Fernández Astudillo,et al. Pushing the Limits of Translation Quality Estimation , 2017, TACL.
[12] Demis Hassabis,et al. Mastering Atari, Go, chess and shogi by planning with a learned model , 2019, Nature.
[13] E. Gumbel. Statistical Theory of Extreme Values and Some Practical Applications : A Series of Lectures , 1954 .
[14] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[15] Mohit Iyyer,et al. Energy-Based Reranking: Improving Neural Machine Translation Using Energy-Based Models , 2020, ACL.
[16] Ning Zhang,et al. Deep Reinforcement Learning-Based Image Captioning with Embedding Reward , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[17] Yejin Choi,et al. The Curious Case of Neural Text Degeneration , 2019, ICLR.
[18] Marc'Aurelio Ranzato,et al. Sequence Level Training with Recurrent Neural Networks , 2015, ICLR.
[19] Geoffrey E. Hinton,et al. A Learning Algorithm for Boltzmann Machines , 1985, Cogn. Sci..
[20] Victor O. K. Li,et al. Non-Autoregressive Neural Machine Translation , 2017, ICLR.
[21] K. Jarrod Millman,et al. Array programming with NumPy , 2020, Nat..
[22] Lucia Specia,et al. Quality Estimation for Machine Translation , 2018, Computational Linguistics.
[23] Anton Osokin,et al. SEARNN: Training RNNs with Global-Local Losses , 2017, ICLR.
[24] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[25] Omri Abend,et al. On the Weaknesses of Reinforcement Learning for Neural Machine Translation , 2019, ICLR.
[26] Marc'Aurelio Ranzato,et al. Classical Structured Prediction Losses for Sequence to Sequence Learning , 2017, NAACL.
[27] Kyunghyun Cho,et al. Non-Monotonic Sequential Text Generation , 2019, ICML.
[28] Noam Shazeer,et al. Fast Transformer Decoding: One Write-Head is All You Need , 2019, ArXiv.
[29] Srivatsan Srinivasan,et al. The DeepMind Chinese–English Document Translation System at WMT2020 , 2020, WMT.
[30] Yu-Siang Wang,et al. Investigating the Decoders of Maximum Likelihood Sequence Models: A Look-ahead Approach , 2020, ArXiv.
[31] Dale Schuurmans,et al. Reward Augmented Maximum Likelihood for Neural Structured Prediction , 2016, NIPS.
[32] Yang Liu,et al. Minimum Risk Training for Neural Machine Translation , 2015, ACL.
[33] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[34] Rémi Coulom,et al. Efficient Selectivity and Backup Operators in Monte-Carlo Tree Search , 2006, Computers and Games.
[35] Wilker Aziz,et al. Is MAP Decoding All You Need? The Inadequacy of the Mode in Neural Machine Translation , 2020, COLING.
[36] Ryan Cotterell,et al. If Beam Search Is the Answer, What Was the Question? , 2020, EMNLP.
[37] Kyunghyun Cho,et al. Consistency of a Recurrent Language Model With Respect to Incomplete Decoding , 2020, EMNLP.
[38] Max Welling,et al. Stochastic Beams and Where to Find Them: The Gumbel-Top-k Trick for Sampling Sequences Without Replacement , 2019, ICML.