暂无分享,去创建一个
[1] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[2] Arman Cohan,et al. Longformer: The Long-Document Transformer , 2020, ArXiv.
[3] Yu Cheng,et al. Diverse Few-Shot Text Classification with Multiple Metrics , 2018, NAACL.
[4] Samuel R. Bowman,et al. Neural Network Acceptability Judgments , 2018, Transactions of the Association for Computational Linguistics.
[5] Chris Brockett,et al. Automatically Constructing a Corpus of Sentential Paraphrases , 2005, IJCNLP.
[6] Oriol Vinyals,et al. Matching Networks for One Shot Learning , 2016, NIPS.
[7] Regina Barzilay,et al. Few-shot Text Classification with Distributional Signatures , 2019, ICLR.
[8] Samuel R. Bowman,et al. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference , 2017, NAACL.
[9] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[10] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[11] Christopher Potts,et al. A large annotated corpus for learning natural language inference , 2015, EMNLP.
[12] Kyunghyun Cho,et al. Mixout: Effective Regularization to Finetune Large-scale Pretrained Language Models , 2020, ICLR.
[13] Hinrich Schutze,et al. It’s Not Just Size That Matters: Small Language Models Are Also Few-Shot Learners , 2020, NAACL.
[14] Cees Snoek,et al. Hyperspherical Prototype Networks , 2019, NeurIPS.
[15] Ido Dagan,et al. The Third PASCAL Recognizing Textual Entailment Challenge , 2007, ACL-PASCAL@ACL.
[16] Andrew M. Dai,et al. Adversarial Training Methods for Semi-Supervised Text Classification , 2016, ICLR.
[17] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[18] Bo Pang,et al. Seeing Stars: Exploiting Class Relationships for Sentiment Categorization with Respect to Rating Scales , 2005, ACL.
[19] Claire Cardie,et al. Annotating Expressions of Opinions and Emotions in Language , 2005, Lang. Resour. Evaluation.
[20] Quoc V. Le,et al. A Simple Method for Commonsense Reasoning , 2018, ArXiv.
[21] Quoc V. Le,et al. Unsupervised Data Augmentation for Consistency Training , 2019, NeurIPS.
[22] Zhiyuan Liu,et al. FewRel: A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation , 2018, EMNLP.
[23] Alexander M. Rush,et al. Commonsense Knowledge Mining from Pretrained Models , 2019, EMNLP.
[24] Bing Liu,et al. Mining and summarizing customer reviews , 2004, KDD.
[25] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[26] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[27] Jonathan Berant,et al. oLMpics-On What Language Model Pre-training Captures , 2019, Transactions of the Association for Computational Linguistics.
[28] Helmut Schmid,et al. Automatically Identifying Words That Can Serve as Labels for Few-Shot Text Classification , 2020, COLING.
[29] Ido Dagan,et al. The Third PASCAL Recognizing Textual Entailment Challenge , 2007, ACL-PASCAL@ACL.
[30] Eneko Agirre,et al. SemEval-2017 Task 1: Semantic Textual Similarity Multilingual and Crosslingual Focused Evaluation , 2017, *SEMEVAL.
[31] Maosong Sun,et al. FewRel 2.0: Towards More Challenging Few-Shot Relation Classification , 2019, EMNLP.
[32] Trapit Bansal,et al. Self-Supervised Meta-Learning for Few-Shot Natural Language Classification Tasks , 2020, EMNLP.
[33] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[34] Roy Bar-Haim,et al. The Second PASCAL Recognising Textual Entailment Challenge , 2006 .
[35] Ellen M. Voorhees,et al. Building a question answering test collection , 2000, SIGIR '00.
[36] Kilian Q. Weinberger,et al. Revisiting Few-sample BERT Fine-tuning , 2020, ArXiv.
[37] Bo Pang,et al. A Sentimental Education: Sentiment Analysis Using Subjectivity Summarization Based on Minimum Cuts , 2004, ACL.
[38] Ido Dagan,et al. The Sixth PASCAL Recognizing Textual Entailment Challenge , 2009, TAC.
[39] Christopher Potts,et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.
[40] Samuel R. Bowman,et al. Sentence Encoders on STILTs: Supplementary Training on Intermediate Labeled-data Tasks , 2018, ArXiv.
[41] Hinrich Schutze,et al. Exploiting Cloze Questions for Few-Shot Text Classification and Natural Language Inference , 2020, ArXiv.
[42] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[43] Ali Farhadi,et al. Fine-Tuning Pretrained Language Models: Weight Initializations, Data Orders, and Early Stopping , 2020, ArXiv.
[44] Andrew McCallum,et al. Learning to Few-Shot Learn Across Diverse Natural Language Classification Tasks , 2020, COLING.
[45] Dragomir R. Radev,et al. Universal Natural Language Processing with Limited Annotations: Try Few-shot Textual Entailment as a Start , 2020, EMNLP.