暂无分享,去创建一个
[1] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[2] Shang-Wen Li,et al. Meta Learning to Classify Intent and Slot Labels with Noisy Few Shot Examples , 2020, 2021 IEEE Spoken Language Technology Workshop (SLT).
[3] James Glass,et al. Cooperative Learning of Zero-Shot Machine Reading Comprehension , 2021, ArXiv.
[4] Dan Roth,et al. A Linear Programming Formulation for Global Inference in Natural Language Tasks , 2004, CoNLL.
[5] Oriol Vinyals,et al. Matching Networks for One Shot Learning , 2016, NIPS.
[6] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[7] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[8] Yuchen Zhang,et al. CoNLL-2012 Shared Task: Modeling Multilingual Unrestricted Coreference in OntoNotes , 2012, EMNLP-CoNLL Shared Task.
[9] Ming-Wei Chang,et al. Natural Questions: A Benchmark for Question Answering Research , 2019, TACL.
[10] Shang-Wen Li,et al. Prototypical Q Networks for Automatic Conversational Diagnosis and Few-Shot New Disease Adaption , 2020, INTERSPEECH.
[11] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[12] Xavier Carreras,et al. Introduction to the CoNLL-2005 Shared Task: Semantic Role Labeling , 2005, CoNLL.
[13] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[14] Omer Levy,et al. SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems , 2019, NeurIPS.
[15] Hao Tian,et al. ERNIE 2.0: A Continual Pre-training Framework for Language Understanding , 2019, AAAI.
[16] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[17] Quoc V. Le,et al. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators , 2020, ICLR.
[18] Richard S. Zemel,et al. Prototypical Networks for Few-shot Learning , 2017, NIPS.