暂无分享,去创建一个
Qun Liu | Xin Jiang | Yi Liao | Xiaoguang Li | Xiao Chen | Wenyong Huang | Yasheng Wang | Junqiu Wei | Xiaozhe Ren | Jiashu Lin
[1] Hao Tian,et al. ERNIE 2.0: A Continual Pre-training Framework for Language Understanding , 2019, AAAI.
[2] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[3] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[4] Hao Wu,et al. Mixed Precision Training , 2017, ICLR.
[5] Yu Sun,et al. ERNIE: Enhanced Representation through Knowledge Integration , 2019, ArXiv.
[6] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[7] Maosong Sun,et al. ERNIE: Enhanced Language Representation with Informative Entities , 2019, ACL.
[8] Guillaume Lample,et al. XNLI: Evaluating Cross-lingual Sentence Representations , 2018, EMNLP.
[9] Wentao Ma,et al. A Span-Extraction Dataset for Chinese Machine Reading Comprehension , 2019, EMNLP-IJCNLP.
[10] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[11] Wanxiang Che,et al. Pre-Training with Whole Word Masking for Chinese BERT , 2019, ArXiv.
[12] Alexander Sergeev,et al. Horovod: fast and easy distributed deep learning in TensorFlow , 2018, ArXiv.
[13] Yiming Yang,et al. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context , 2019, ACL.
[14] Yu Sun,et al. ERNIE 2 . 0 : A CONTINUAL PRE-TRAINING FRAMEWORK FOR LANGUAGE UNDERSTANDING , 2019 .
[15] Xin Liu,et al. LCQMC:A Large-scale Chinese Question Matching Corpus , 2018, COLING.
[16] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[17] James Demmel,et al. Reducing BERT Pre-Training Time from 3 Days to 76 Minutes , 2019, ArXiv.
[18] Ashish Vaswani,et al. Self-Attention with Relative Position Representations , 2018, NAACL.
[19] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.