暂无分享,去创建一个
[1] Qiang Yang,et al. A Survey on Transfer Learning , 2010, IEEE Transactions on Knowledge and Data Engineering.
[2] Ji Wang,et al. Pretraining-Based Natural Language Generation for Text Summarization , 2019, CoNLL.
[3] Mirella Lapata,et al. Neural Latent Extractive Document Summarization , 2018, EMNLP.
[4] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[5] Phil Blunsom,et al. Teaching Machines to Read and Comprehend , 2015, NIPS.
[6] Richard Socher,et al. A Deep Reinforced Model for Abstractive Summarization , 2017, ICLR.
[7] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[8] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[9] Ronald J. Williams,et al. A Learning Algorithm for Continually Running Fully Recurrent Neural Networks , 1989, Neural Computation.
[10] Jiusheng Chen,et al. ProphetNet: Predicting Future N-gram for Sequence-to-Sequence Pre-training , 2020, EMNLP.
[11] Ido Dagan,et al. Better Rewards Yield Better Summaries: Learning to Summarise Without References , 2019, EMNLP.
[12] Yen-Chun Chen,et al. Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting , 2018, ACL.
[13] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[14] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[15] Jihoon Kim,et al. Summary Level Training of Sentence Rewriting for Abstractive Summarization , 2019, EMNLP.
[16] Christopher D. Manning,et al. Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.
[17] Mirella Lapata,et al. Ranking Sentences for Extractive Summarization with Reinforcement Learning , 2018, NAACL.
[18] Alexander M. Rush,et al. Bottom-Up Abstractive Summarization , 2018, EMNLP.
[19] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[20] Glen Jeh. Encoder-decoder Network as Loss Function for Summarization , 2019 .
[21] Ramakanth Pasunuru,et al. Multi-Reward Reinforced Summarization with Saliency and Entailment , 2018, NAACL.
[22] Yao Zhao,et al. PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization , 2020, ICML.
[23] Zhe Gan,et al. Adversarial Text Generation via Feature-Mover's Distance , 2018, NeurIPS.
[24] Pat Langley,et al. Crafting Papers on Machine Learning , 2000, ICML.
[25] Weijia Jia,et al. Improving Abstractive Document Summarization with Salient Information Modeling , 2019, ACL.
[26] Mirella Lapata,et al. Text Summarization with Pretrained Encoders , 2019, EMNLP.
[27] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[28] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.