暂无分享,去创建一个
[1] Matthew Henderson,et al. ConveRT: Efficient and Accurate Conversational Representations from Transformers , 2020, EMNLP.
[2] Quoc V. Le,et al. Self-Training With Noisy Student Improves ImageNet Classification , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[3] Pieter Abbeel,et al. Emergence of Grounded Compositional Language in Multi-Agent Populations , 2017, AAAI.
[4] Alexander Peysakhovich,et al. Multi-Agent Cooperation and the Emergence of (Natural) Language , 2016, ICLR.
[5] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[6] Guy Lever,et al. Deterministic Policy Gradient Algorithms , 2014, ICML.
[7] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[8] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[9] Vaibhava Goel,et al. Self-Critical Sequence Training for Image Captioning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[10] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[11] Ramesh Nallapati,et al. End-to-End Synthetic Data Generation for Domain Adaptation of Question Answering Systems , 2020, EMNLP.
[12] Yoshua Bengio,et al. Straight to the Tree: Constituency Parsing with Neural Syntactic Distance , 2018, ACL.
[13] Thomas Wolf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[14] G P Shrivatsa Bhargav,et al. Span Selection Pre-training for Question Answering , 2019, ACL.
[15] Dong Bok Lee,et al. Generating Diverse and Consistent QA pairs from Contexts with Information-Maximizing Hierarchical Conditional VAEs , 2020, ACL.
[16] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[17] Omer Levy,et al. SpanBERT: Improving Pre-training by Representing and Predicting Spans , 2019, TACL.
[18] Jason Weston,et al. Real-time Inference in Multi-sentence Tasks with Deep Pretrained Transformers , 2019, ArXiv.
[19] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[20] Sham M. Kakade,et al. A Natural Policy Gradient , 2001, NIPS.
[21] Ming-Wei Chang,et al. Natural Questions: A Benchmark for Question Answering Research , 2019, TACL.
[22] Donghong Ji,et al. Retrofitting Structure-aware Transformer Language Model for End Tasks , 2020, EMNLP.
[23] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[24] Jason Weston,et al. Poly-encoders: Transformer Architectures and Pre-training Strategies for Fast and Accurate Multi-sentence Scoring , 2019 .
[25] Ian J. Goodfellow,et al. NIPS 2016 Tutorial: Generative Adversarial Networks , 2016, ArXiv.
[26] Monica S. Lam,et al. AutoQA: From Databases to Q&A Semantic Parsers with Only Synthetic Training Data , 2020, EMNLP.
[27] Aaron C. Courville,et al. StructFormer: Joint Unsupervised Induction of Dependency and Constituency Structure from Masked Language Modeling , 2020, ACL.
[28] Jiajun Shen,et al. Revisiting Self-Training for Neural Sequence Generation , 2020, ICLR.
[29] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[30] Ivan Titov,et al. Emergence of Language with Multi-agent Games: Learning to Communicate with Sequences of Symbols , 2017, NIPS.
[31] Yonatan Belinkov,et al. Improving Neural Language Models by Segmenting, Attending, and Predicting the Future , 2019, ACL.
[32] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[33] Hermann Ney,et al. Maximum Entropy Models for Named Entity Recognition , 2003, CoNLL.
[34] Danqi Chen,et al. Dense Passage Retrieval for Open-Domain Question Answering , 2020, EMNLP.
[35] Quoc V. Le,et al. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators , 2020, ICLR.
[36] Lantao Yu,et al. SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient , 2016, AAAI.
[37] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[38] Sebastian Riedel,et al. Beat the AI: Investigating Adversarial Human Annotation for Reading Comprehension , 2020, Transactions of the Association for Computational Linguistics.