Improving Language Understanding by Generative Pre-Training
暂无分享,去创建一个
[1] Percy Liang,et al. Semi-Supervised Learning for Natural Language , 2005 .
[2] Chris Brockett,et al. Automatically Constructing a Corpus of Sentential Paraphrases , 2005, IJCNLP.
[3] Xiaojin Zhu,et al. Semi-Supervised Learning Literature Survey , 2005 .
[4] Yee Whye Teh,et al. A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.
[5] Tom M. Mitchell,et al. Semi-Supervised Text Classification Using EM , 2006, Semi-Supervised Learning.
[6] Marc'Aurelio Ranzato,et al. Efficient Learning of Sparse Representations with an Energy-Based Model , 2006, NIPS.
[7] Dale Schuurmans,et al. Semi-Supervised Conditional Random Fields for Improved Sequence Segmentation and Labeling , 2006, ACL.
[8] H. Robbins. A Stochastic Approximation Method , 1951 .
[9] Thomas Hofmann,et al. Greedy Layer-Wise Training of Deep Networks , 2007 .
[10] Jason Weston,et al. A unified architecture for natural language processing: deep neural networks with multitask learning , 2008, ICML '08.
[11] Jun Suzuki,et al. Semi-Supervised Sequential Labeling and Segmentation Using Giga-Word Scale Unlabeled Data , 2008, ACL.
[12] Yoshua Bengio,et al. Extracting and composing robust features with denoising autoencoders , 2008, ICML '08.
[13] Ido Dagan,et al. The Sixth PASCAL Recognizing Textual Entailment Challenge , 2009, TAC.
[14] Dong Yu,et al. Roles of Pre-Training and Fine-Tuning in Context-Dependent DBN-HMMs for Real-World Speech Recognition , 2010 .
[15] Jason Weston,et al. Natural Language Processing (Almost) from Scratch , 2011, J. Mach. Learn. Res..
[16] Vincent Ng,et al. Resolving Complex Cases of Definite Pronouns: The Winograd Schema Challenge , 2012, EMNLP.
[17] Jacob Eisenstein,et al. Discriminative Improvements to Distributional Sentence Similarity , 2013, EMNLP.
[18] Houfeng Wang,et al. Learning Entity Representation for Entity Disambiguation , 2013, ACL.
[19] Christopher Potts,et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.
[20] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[21] Peter Glöckner,et al. Why Does Unsupervised Pre-training Help Deep Learning? , 2013 .
[22] Yoon Kim,et al. Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.
[23] Danqi Chen,et al. A Fast and Accurate Dependency Parser using Neural Networks , 2014, EMNLP.
[24] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[25] Quoc V. Le,et al. Distributed Representations of Sentences and Documents , 2014, ICML.
[26] Sanja Fidler,et al. Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[27] Quoc V. Le,et al. Semi-supervised Sequence Learning , 2015, NIPS.
[28] Sanja Fidler,et al. Skip-Thought Vectors , 2015, NIPS.
[29] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[30] Phil Blunsom,et al. Teaching Machines to Read and Comprehend , 2015, NIPS.
[31] Christopher Potts,et al. A large annotated corpus for learning natural language inference , 2015, EMNLP.
[32] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[33] Phil Blunsom,et al. Reasoning about Entailment with Neural Attention , 2015, ICLR.
[34] Kevin Gimpel,et al. Bridging Nonlinearities and Stochastic Regularizers with Gaussian Error Linear Units , 2016, ArXiv.
[35] Geoffrey E. Hinton,et al. Layer Normalization , 2016, ArXiv.
[36] Chandra Bhagavatula,et al. Semi-supervised sequence tagging with bidirectional language models , 2017, ACL.
[37] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[38] Dan Roth,et al. Story Comprehension for Predicting What Happens Next , 2017, EMNLP.
[39] Sanjeev Arora,et al. A Simple but Tough-to-Beat Baseline for Sentence Embeddings , 2017, ICLR.
[40] Frank Hutter,et al. Fixing Weight Decay Regularization in Adam , 2017, ArXiv.
[41] Xiaodong Liu,et al. Towards Human-level Machine Reading Comprehension: Reasoning and Inference with Multiple Strategies , 2017, ArXiv.
[42] Piotr,et al. UNSUPERVISED MACHINE TRANSLATION USING MONOLINGUAL CORPORA ONLY , 2017 .
[43] Alexei A. Efros,et al. Split-Brain Autoencoders: Unsupervised Learning by Cross-Channel Prediction , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[44] Guokun Lai,et al. RACE: Large-scale ReAding Comprehension Dataset From Examinations , 2017, EMNLP.
[45] Quoc V. Le,et al. Unsupervised Pretraining for Sequence to Sequence Learning , 2016, EMNLP.
[46] Hongbo Zhang,et al. Quora Question Pairs , 2017 .
[47] Nathanael Chambers,et al. LSDSem 2017 Shared Task: The Story Cloze Test , 2017, LSDSem@EACL.
[48] Diederik P. Kingma,et al. GPU Kernels for Block-Sparse Weights , 2017 .
[49] Eneko Agirre,et al. SemEval-2017 Task 1: Semantic Textual Similarity Multilingual and Crosslingual Focused Evaluation , 2017, *SEMEVAL.
[50] Samuel R. Bowman,et al. Discourse-Based Objectives for Fast Unsupervised Sentence Representation Learning , 2017, ArXiv.
[51] Marek Rei,et al. Semi-supervised Multitask Learning for Sequence Labeling , 2017, ACL.
[52] Richard Socher,et al. Learned in Translation: Contextualized Word Vectors , 2017, NIPS.
[53] Man Lan,et al. ECNU at SemEval-2017 Task 1: Leverage Kernel-based Traditional NLP features and Neural Networks to Build a Universal Model for Multilingual and Cross-lingual Semantic Textual Similarity , 2017, SemEval@ACL.
[54] Holger Schwenk,et al. Supervised Learning of Universal Sentence Representations from Natural Language Inference Data , 2017, EMNLP.
[55] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[56] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[57] Samuel R. Bowman,et al. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference , 2017, NAACL.
[58] Xiaodong Liu,et al. Stochastic Answer Networks for Natural Language Inference , 2018, ArXiv.
[59] Mark O. Riedl,et al. A Simple and Effective Approach to the Story Cloze Test , 2018, NAACL-HLT.
[60] Lukasz Kaiser,et al. Generating Wikipedia by Summarizing Long Sequences , 2018, ICLR.
[61] Dan Klein,et al. Constituency Parsing with a Self-Attentive Encoder , 2018, ACL.
[62] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[63] Christopher Joseph Pal,et al. Learning General Purpose Distributed Sentence Representations via Large Scale Multi-task Learning , 2018, ICLR.
[64] Honglak Lee,et al. An efficient framework for learning sentence representations , 2018, ICLR.
[65] Peter Clark,et al. SciTaiL: A Textual Entailment Dataset from Science Question Answering , 2018, AAAI.
[66] Siu Cheung Hui,et al. Multi-range Reasoning for Machine Comprehension , 2018, ArXiv.
[67] Guillaume Lample,et al. Unsupervised Machine Translation Using Monolingual Corpora Only , 2017, ICLR.
[68] Siu Cheung Hui,et al. A Compare-Propagate Architecture with Alignment Factorization for Natural Language Inference , 2017, ArXiv.