Recurrent Inference in Text Editing
暂无分享,去创建一个
[1] Oren Etzioni,et al. Learning to Solve Arithmetic Word Problems with Verb Categorization , 2014, EMNLP.
[2] Lutz Prechelt,et al. Early Stopping - But When? , 2012, Neural Networks: Tricks of the Trade.
[3] Olivier Pietquin,et al. LIG-CRIStAL Submission for the WMT 2017 Automatic Post-Editing Task , 2017, WMT.
[4] Dan Roth,et al. Solving General Arithmetic Word Problems , 2016, EMNLP.
[5] Shashi Narayan,et al. Hybrid Simplification using Deep Semantics and Machine Translation , 2014, ACL.
[6] Nando de Freitas,et al. Neural Programmer-Interpreters , 2015, ICLR.
[7] Joachim Bingel,et al. Learning How to Simplify From Explicit Labeling of Complex-Simplified Text Pairs , 2017, IJCNLP.
[8] Tanel Alumäe,et al. Bidirectional Recurrent Neural Network with Attention Mechanism for Punctuation Restoration , 2016, INTERSPEECH.
[9] Ming Zhou,et al. Fluency Boost Learning and Inference for Neural Grammatical Error Correction , 2018, ACL.
[10] Sepp Hochreiter,et al. The Vanishing Gradient Problem During Learning Recurrent Neural Nets and Problem Solutions , 1998, Int. J. Uncertain. Fuzziness Knowl. Based Syst..
[11] Marco Baroni,et al. Generalization without Systematicity: On the Compositional Skills of Sequence-to-Sequence Recurrent Networks , 2017, ICML.
[12] Armando Solar-Lezama,et al. Write, Execute, Assess: Program Synthesis with a REPL , 2019, NeurIPS.
[13] Lejian Liao,et al. Can Syntax Help? Improving an LSTM-based Sentence Compression Model for New Domains , 2017, ACL.
[14] Christoph Meinel,et al. Punctuation Prediction for Unsegmented Transcript Based on Word Vector , 2016, LREC.
[15] Yoshua Bengio,et al. On the Properties of Neural Machine Translation: Encoder–Decoder Approaches , 2014, SSST@EMNLP.
[16] Aliaksei Severyn,et al. Encode, Tag, Realize: High-Precision Text Editing , 2019, EMNLP.
[17] Wei Zhao,et al. Improving Grammatical Error Correction via Pre-Training a Copy-Augmented Architecture with Unlabeled Data , 2019, NAACL.
[18] Gholamreza Haffari,et al. Automatic Post-Editing of Machine Translation: A Neural Programmer-Interpreter Approach , 2018, EMNLP.
[19] Razvan Pascanu,et al. On the difficulty of training recurrent neural networks , 2012, ICML.
[20] Christopher D. Manning,et al. Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.
[21] Kuldip K. Paliwal,et al. Bidirectional recurrent neural networks , 1997, IEEE Trans. Signal Process..
[22] Seokhwan Kim,et al. Deep Recurrent Neural Networks with Layer-wise Multi-head Attentions for Punctuation Restoration , 2019, ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[23] Wojciech Zaremba,et al. Learning to Execute , 2014, ArXiv.
[24] Xiaojun Wan,et al. Abstractive Document Summarization with a Graph-Based Attentional Neural Model , 2017, ACL.
[25] Mirella Lapata,et al. Sentence Simplification with Deep Reinforcement Learning , 2017, EMNLP.
[26] Wang Ling,et al. Program Induction by Rationale Generation: Learning to Solve and Explain Algebraic Word Problems , 2017, ACL.
[27] Navdeep Jaitly,et al. Pointer Networks , 2015, NIPS.
[28] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[29] Jackie Chi Kit Cheung,et al. EditNTS: An Neural Programmer-Interpreter Model for Sentence Simplification through Explicit Editing , 2019, ACL.
[30] Guillaume Lample,et al. Deep Learning for Symbolic Mathematics , 2019, ICLR.
[31] Pushmeet Kohli,et al. Analysing Mathematical Reasoning Abilities of Neural Models , 2019, ICLR.
[32] Lukasz Kaiser,et al. Sentence Compression by Deletion with LSTMs , 2015, EMNLP.
[33] Sergiu Nisioi,et al. Exploring Neural Text Simplification Models , 2017, ACL.
[34] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[35] Alexander M. Rush,et al. Learning Neural Templates for Text Generation , 2018, EMNLP.
[36] Pushmeet Kohli,et al. RobustFill: Neural Program Learning under Noisy I/O , 2017, ICML.
[37] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[38] Christopher D. Manning,et al. Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.
[39] Changhan Wang,et al. Levenshtein Transformer , 2019, NeurIPS.
[40] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[41] Guigang Zhang,et al. Deep Learning , 2016, Int. J. Semantic Comput..
[42] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[43] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[44] Shankar Kumar,et al. Weakly Supervised Grammatical Error Correction using Iterative Decoding , 2018, ArXiv.
[45] Samuel R. Bowman,et al. ListOps: A Diagnostic Dataset for Latent Tree Learning , 2018, NAACL.
[46] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[47] Marcin Junczys-Dowmunt,et al. Approaching Neural Grammatical Error Correction as a Low-Resource Machine Translation Task , 2018, NAACL.
[48] Mohammad Norouzi,et al. The Importance of Generation Order in Language Modeling , 2018, EMNLP.
[49] Jindřich Helcl,et al. CUNI System for WMT16 Automatic Post-Editing and Multimodal Translation Tasks , 2016, WMT.