暂无分享,去创建一个
[1] Ruslan Salakhutdinov,et al. Breaking the Softmax Bottleneck: A High-Rank RNN Language Model , 2017, ICLR.
[2] Hwee Tou Ng,et al. Better Evaluation for Grammatical Error Correction , 2012, NAACL.
[3] Myle Ott,et al. fairseq: A Fast, Extensible Toolkit for Sequence Modeling , 2019, NAACL.
[4] Samy Bengio,et al. Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks , 2015, NIPS.
[5] Dan Klein,et al. Train Large, Then Compress: Rethinking Model Size for Efficient Training and Inference of Transformers , 2020, ArXiv.
[6] Ted Briscoe,et al. Automatic Extraction of Learner Errors in ESL Sentences Using Linguistically Enhanced Alignments , 2016, COLING.
[7] Masaaki Nagata,et al. Direct Output Connection for a High-Rank Language Model , 2018, EMNLP.
[8] Andrew M. Dai,et al. Adversarial Training Methods for Semi-Supervised Text Classification , 2016, ICLR.
[9] Matt Post,et al. A Call for Clarity in Reporting BLEU Scores , 2018, WMT.
[10] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[11] Rico Sennrich,et al. Improving Neural Machine Translation Models with Monolingual Data , 2015, ACL.
[12] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[13] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[14] Ming Zhou,et al. ProphetNet: Predicting Future N-gram for Sequence-to-Sequence Pre-training , 2020, FINDINGS.
[15] Raymond Hendy Susanto,et al. The CoNLL-2014 Shared Task on Grammatical Error Correction , 2014 .
[16] Xu Tan,et al. MASS: Masked Sequence to Sequence Pre-training for Language Generation , 2019, ICML.
[17] Ted Briscoe,et al. Automatic Annotation and Evaluation of Error Types for Grammatical Error Correction , 2017, ACL.
[18] Jason Weston,et al. A Neural Attention Model for Abstractive Sentence Summarization , 2015, EMNLP.
[19] Benjamin Van Durme,et al. Annotated Gigaword , 2012, AKBC-WEKEX@NAACL-HLT.
[20] Shin Ishii,et al. Distributional Smoothing with Virtual Adversarial Training , 2015, ICLR 2016.
[21] Myle Ott,et al. Scaling Neural Machine Translation , 2018, WMT.
[22] Rico Sennrich,et al. Revisiting Low-Resource Neural Machine Translation: A Case Study , 2019, ACL.
[23] Sosuke Kobayashi,et al. Contextual Augmentation: Data Augmentation by Words with Paradigmatic Relations , 2018, NAACL.
[24] Richard Socher,et al. Regularizing and Optimizing LSTM Language Models , 2017, ICLR.
[25] Daniel Jurafsky,et al. Data Noising as Smoothing in Neural Network Language Models , 2017, ICLR.
[26] Zoubin Ghahramani,et al. A Theoretically Grounded Application of Dropout in Recurrent Neural Networks , 2015, NIPS.
[27] Marta R. Costa-jussà,et al. Findings of the 2019 Conference on Machine Translation (WMT19) , 2019, WMT.
[28] Andrew McCallum,et al. Energy and Policy Considerations for Deep Learning in NLP , 2019, ACL.
[29] Oren Etzioni,et al. Green AI , 2019, Commun. ACM.
[30] Dilin Wang,et al. Improving Neural Language Modeling via Adversarial Training , 2019, ICML.
[31] Ted Briscoe,et al. The BEA-2019 Shared Task on Grammatical Error Correction , 2019, BEA@ACL.
[32] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[33] Yang Feng,et al. Bridging the Gap between Training and Inference for Neural Machine Translation , 2019, ACL.
[34] Jianfeng Gao,et al. A Nested Attention Neural Hybrid Model for Grammatical Error Correction , 2017, ACL.
[35] Ali Farhadi,et al. Defending Against Neural Fake News , 2019, NeurIPS.
[36] Xiaodong Liu,et al. Unified Language Model Pre-training for Natural Language Understanding and Generation , 2019, NeurIPS.
[37] Yong Cheng,et al. Robust Neural Machine Translation with Doubly Adversarial Inputs , 2019, ACL.
[38] Jun Suzuki,et al. Effective Adversarial Regularization for Neural Machine Translation , 2019, ACL.
[39] Yao Zhao,et al. PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization , 2020, ICML.
[40] Masaaki Nagata,et al. Cutting-off Redundant Repeating Generations for Neural Abstractive Summarization , 2016, EACL.
[41] Samy Bengio,et al. Parallel Scheduled Sampling , 2019, ArXiv.
[42] Ciprian Chelba,et al. Tagged Back-Translation , 2019, WMT.
[43] Naoaki Okazaki,et al. Positional Encoding to Control Output Sequence Length , 2019, NAACL.
[44] Jun Suzuki,et al. Massive Exploration of Pseudo Data for Grammatical Error Correction , 2020, IEEE/ACM Transactions on Audio, Speech, and Language Processing.
[45] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[46] Sosuke Kobayashi,et al. All Word Embeddings from One Embedding , 2020, NeurIPS.
[47] Paul Over,et al. DUC in context , 2007, Inf. Process. Manag..