Co-Teaching Student-Model through Submission Results of Shared Task
暂无分享,去创建一个
Yukino Baba | Satoshi Sekine | Shuhei Kurita | Kouta Nakayama | Akio Kobayashi | S. Sekine | Yukino Baba | Shuhei Kurita | Kouta Nakayama | Akio Kobayashi
[1] Erik F. Tjong Kim Sang,et al. Representing Text Chunks , 1999, EACL.
[2] Preslav Nakov,et al. SemEval-2010 Task 8: Multi-Way Classification of Semantic Relations Between Pairs of Nominals , 2009, SEW@NAACL-HLT.
[3] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[4] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[5] Jason Weston,et al. Reading Wikipedia to Answer Open-Domain Questions , 2017, ACL.
[6] Erik F. Tjong Kim Sang,et al. Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition , 2003, CoNLL.
[7] Kevin Gimpel,et al. Gaussian Error Linear Units (GELUs) , 2016 .
[8] Satoshi Sekine,et al. Extended Named Entity Ontology with Attribute Information , 2008, LREC.
[9] Avrim Blum,et al. The Bottleneck , 2021, Monopsony Capitalism.
[10] Vladlen Koltun,et al. Multi-Task Learning as Multi-Objective Optimization , 2018, NeurIPS.
[11] Hoa Trang Dang,et al. Overview of the TAC 2008 Update Summarization Task , 2008, TAC.
[12] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[13] Iadh Ounis,et al. Overview of the TREC 2008 Blog Track , 2008, TREC.
[14] Naoaki Okazaki,et al. A Joint Neural Model for Fine-Grained Named Entity Classification of Wikipedia Articles , 2018, IEICE Trans. Inf. Syst..
[15] Nadir Durrani,et al. FINDINGS OF THE IWSLT 2020 EVALUATION CAMPAIGN , 2020, IWSLT.
[16] Yang Song,et al. Class-Balanced Loss Based on Effective Number of Samples , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[17] Houda Bouamor,et al. NADI 2021: The Second Nuanced Arabic Dialect Identification Shared Task , 2021, WANLP.
[18] Akio Kobayashi,et al. SHINRA: Structuring Wikipedia by Collaborative Contribution , 2019, AKBC.
[19] Isao Goto,et al. Overview of the 7th Workshop on Asian Translation , 2020, WAT.
[20] Ross B. Girshick,et al. Focal Loss for Dense Object Detection , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[21] Benno Stein,et al. TIRA Integrated Research Architecture , 2019, Information Retrieval Evaluation in a Changing World.
[22] David Yarowsky,et al. Unsupervised Word Sense Disambiguation Rivaling Supervised Methods , 1995, ACL.
[23] Sabine Buchholz,et al. Introduction to the CoNLL-2000 Shared Task Chunking , 2000, CoNLL/LLL.
[24] Eduard Hovy,et al. Overview and Insights from the Shared Tasks at Scholarly Document Processing 2020: CL-SciSumm, LaySumm and LongSumm , 2020, SDP.
[25] Yoshua Bengio,et al. FitNets: Hints for Thin Deep Nets , 2014, ICLR.
[26] Rich Caruana,et al. Do Deep Nets Really Need to be Deep? , 2013, NIPS.
[27] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[28] George Giannakopoulos,et al. The Financial Narrative Summarisation Shared Task (FNS 2020) , 2020, FNP.
[29] Philipp Koehn,et al. Findings of the 2020 Conference on Machine Translation (WMT20) , 2020, WMT.
[30] Huchuan Lu,et al. Deep Mutual Learning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[31] Beth M. Sundheim,et al. Overview of Results of the MUC-6 Evaluation , 1995, MUC.