XLNet: Generalized Autoregressive Pretraining for Language Understanding
暂无分享,去创建一个
Yiming Yang | Quoc V. Le | Ruslan Salakhutdinov | Zhilin Yang | Zihang Dai | Jaime Carbonell | R. Salakhutdinov | Zhilin Yang | J. Carbonell | Yiming Yang | Zihang Dai
[1] Samy Bengio,et al. Modeling High-Dimensional Discrete Data with Multi-Layer Neural Networks , 1999, NIPS.
[2] Sanja Fidler,et al. Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[3] Quoc V. Le,et al. Semi-supervised Sequence Learning , 2015, NIPS.
[4] Xiang Zhang,et al. Character-level Convolutional Networks for Text Classification , 2015, NIPS.
[5] Hugo Larochelle,et al. MADE: Masked Autoencoder for Distribution Estimation , 2015, ICML.
[6] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[7] Hugo Larochelle,et al. Neural Autoregressive Distribution Estimation , 2016, J. Mach. Learn. Res..
[8] Koray Kavukcuoglu,et al. Pixel Recurrent Neural Networks , 2016, ICML.
[9] Andrew M. Dai,et al. Virtual Adversarial Training for Semi-Supervised Text Classification , 2016, ArXiv.
[10] W. Bruce Croft,et al. A Deep Relevance Matching Model for Ad-hoc Retrieval , 2016, CIKM.
[11] Christopher Ré,et al. Snorkel: Rapid Training Data Creation with Weak Supervision , 2017, Proc. VLDB Endow..
[12] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[13] Andrew M. Dai,et al. Adversarial Training Methods for Semi-Supervised Text Classification , 2016, ICLR.
[14] Tie-Yan Liu,et al. Word-Entity Duet Representations for Document Ranking , 2017, SIGIR.
[15] Philip Bachman,et al. NewsQA: A Machine Comprehension Dataset , 2016, Rep4NLP@ACL.
[16] Guokun Lai,et al. RACE: Large-scale ReAding Comprehension Dataset From Examinations , 2017, EMNLP.
[17] Tong Zhang,et al. Deep Pyramid Convolutional Neural Networks for Text Categorization , 2017, ACL.
[18] Richard Socher,et al. Learned in Translation: Contextualized Word Vectors , 2017, NIPS.
[19] Zhiyuan Liu,et al. End-to-End Neural Ad-hoc Ranking with Kernel Pooling , 2017, SIGIR.
[20] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[21] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[22] Percy Liang,et al. Know What You Don’t Know: Unanswerable Questions for SQuAD , 2018, ACL.
[23] Andrew M. Dai,et al. MaskGAN: Better Text Generation via Filling in the ______ , 2018, ICLR.
[24] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[25] Taku Kudo,et al. SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing , 2018, EMNLP.
[26] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[27] Ruslan Salakhutdinov,et al. Breaking the Softmax Bottleneck: A High-Rank RNN Language Model , 2017, ICLR.
[28] Zhiyuan Liu,et al. Convolutional Neural Networks for Soft-Matching N-Grams in Ad-hoc Search , 2018, WSDM.
[29] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[30] Quoc V. Le,et al. Unsupervised Data Augmentation , 2019, ArXiv.
[31] Xiaodong Liu,et al. Multi-Task Deep Neural Networks for Natural Language Understanding , 2019, ACL.
[32] Heng Ji,et al. Improving Question Answering with External Knowledge , 2019, EMNLP.
[33] Noah Constant,et al. Character-Level Language Modeling with Deeper Self-Attention , 2018, AAAI.
[34] Ruslan Salakhutdinov,et al. Revisiting LSTM Networks for Semi-Supervised Text Classification via Mixed Objective Function , 2019, AAAI.
[35] Yiming Yang,et al. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context , 2019, ACL.
[36] Thomas Lukasiewicz,et al. A Surprisingly Robust Trick for the Winograd Schema Challenge , 2019, ACL.
[37] Alexei Baevski,et al. Adaptive Input Representations for Neural Language Modeling , 2018, ICLR.
[38] Peng Li,et al. Option Comparison Network for Multiple-choice Reading Comprehension , 2019, ArXiv.
[39] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[40] Quoc V. Le,et al. BAM! Born-Again Multi-Task Networks for Natural Language Understanding , 2019, ACL.
[41] Kevin Gimpel,et al. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations , 2019, ICLR.
[42] Xiaoping Zhou,et al. Dual Co-Matching Network for Multi-choice Reading Comprehension , 2019, AAAI.