暂无分享,去创建一个
Maosong Sun | Ning Ding | Hai-Tao Zheng | Shengding Hu | Yulin Chen | Zhiyuan Liu | Weilin Zhao | Haitao Zheng | Maosong Sun | Zhiyuan Liu | Shengding Hu | Ning Ding | Weilin Zhao | Yulin Chen
[1] Samuel R. Bowman,et al. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference , 2017, NAACL.
[2] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[3] Fabio Petroni,et al. Cutting Down on Prompts and Parameters: Simple Few-Shot Learning with Language Models , 2021, FINDINGS.
[4] Zhengxiao Du,et al. GPT Understands, Too , 2021, AI Open.
[5] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[6] Claire Gardent,et al. The WebNLG Challenge: Generating Text from RDF Data , 2017, INLG.
[7] Zhiyuan Liu,et al. PTR: Prompt Tuning with Rules for Text Classification , 2021, AI Open.
[8] Ning Ding,et al. Prompt-Learning for Fine-Grained Entity Typing , 2021, EMNLP.
[9] Xu Tan,et al. MASS: Masked Sequence to Sequence Pre-training for Language Generation , 2019, ICML.
[10] Christopher Potts,et al. Learning Word Vectors for Sentiment Analysis , 2011, ACL.
[11] Ian S. Dunn,et al. Exploring the Limits , 2009 .
[12] Timo Schick,et al. Exploiting Cloze-Questions for Few-Shot Text Classification and Natural Language Inference , 2020, EACL.
[13] Brian Lester,et al. The Power of Scale for Parameter-Efficient Prompt Tuning , 2021, EMNLP.
[14] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[15] Jens Lehmann,et al. DBpedia - A large-scale, multilingual knowledge base extracted from Wikipedia , 2015, Semantic Web.
[16] Lysandre Debut,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[17] Preslav Nakov,et al. SemEval-2010 Task 8: Multi-Way Classification of Semantic Relations Between Pairs of Nominals , 2009, SEW@NAACL-HLT.
[18] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[19] Minlie Huang,et al. PPT: Pre-trained Prompt Tuning for Few-shot Learning , 2021, ArXiv.
[20] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[21] Brian Lester,et al. SPoT: Better Frozen Model Adaptation through Soft Prompt Transfer , 2021, ACL.
[22] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[23] Zhiyuan Liu,et al. Few-NERD: A Few-shot Named Entity Recognition Dataset , 2021, ACL.
[24] Maosong Sun,et al. Exploring Low-dimensional Intrinsic Task Subspace via Prompt Tuning , 2021, ArXiv.
[25] Xipeng Qiu,et al. Pre-trained models for natural language processing: A survey , 2020, Science China Technological Sciences.
[26] Zhiyuan Liu,et al. Pre-Trained Models: Past, Present and Future , 2021, AI Open.
[27] Percy Liang,et al. Prefix-Tuning: Optimizing Continuous Prompts for Generation , 2021, ACL.
[28] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[29] Omer Levy,et al. SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems , 2019, NeurIPS.
[30] Xiang Zhang,et al. Character-level Convolutional Networks for Text Classification , 2015, NIPS.
[31] Colin Wei,et al. Why Do Pretrained Language Models Help in Downstream Tasks? An Analysis of Head and Prompt Tuning , 2021, ArXiv.
[32] Maosong Sun,et al. Knowledgeable Prompt-tuning: Incorporating Knowledge into Prompt Verbalizer for Text Classification , 2021, ArXiv.
[33] Danqi Chen,et al. Making Pre-trained Language Models Better Few-shot Learners , 2021, ACL/IJCNLP.
[34] Dan Klein,et al. Calibrate Before Use: Improving Few-Shot Performance of Language Models , 2021, ICML.
[35] Alexander M. Rush,et al. How many data points is a prompt worth? , 2021, NAACL.