暂无分享,去创建一个
[1] Daniel Jurafsky,et al. Sharp Nearby, Fuzzy Far Away: How Neural Language Models Use Context , 2018, ACL.
[2] Sebastian Ruder,et al. Fine-tuned Language Models for Text Classification , 2018, ArXiv.
[3] Jason Baldridge,et al. Mind the GAP: A Balanced Corpus of Gendered Ambiguous Pronouns , 2018, TACL.
[4] Trevor Hastie,et al. The Elements of Statistical Learning , 2001 .
[5] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[6] Hwee Tou Ng,et al. A Nil-Aware Answer Extraction Framework for Question Answering , 2018, EMNLP.
[7] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[8] Luke S. Zettlemoyer,et al. Higher-Order Coreference Resolution with Coarse-to-Fine Inference , 2018, NAACL.
[9] Richard Socher,et al. Coarse-grain Fine-grain Coattention Network for Multi-evidence Question Answering , 2019, ICLR.
[10] Luke S. Zettlemoyer,et al. The Referential Reader: A Recurrent Entity Network for Anaphora Resolution , 2019, ACL.
[11] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[12] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[13] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[14] Yiming Yang,et al. Transformer-XL: Language Modeling with Longer-Term Dependency , 2018 .
[15] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[16] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.