暂无分享,去创建一个
[1] R'emi Louf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[2] Mirella Lapata,et al. Text Summarization with Pretrained Encoders , 2019, EMNLP.
[3] Derek Miller,et al. Leveraging BERT for Extractive Text Summarization on Lectures , 2019, ArXiv.
[4] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[5] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[6] Jesse Vig,et al. A Multiscale Visualization of Attention in the Transformer Model , 2019, ACL.
[7] Yejin Choi,et al. The Curious Case of Neural Text Degeneration , 2019, ICLR.
[8] Wenpeng Yin,et al. Empirical evaluation of multi-task learning in deep neural networks for natural language processing , 2020, Neural Computing and Applications.
[9] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[10] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[11] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[12] Oren Etzioni,et al. CORD-19: The Covid-19 Open Research Dataset , 2020, NLPCOVID19.
[13] Thomas Wolf,et al. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter , 2019, ArXiv.
[14] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[15] Naren Ramakrishnan,et al. Neural Abstractive Text Summarization with Sequence-to-Sequence Models , 2018, Trans. Data Sci..