暂无分享,去创建一个
Junji Tomita | Kyosuke Nishida | Itsumi Saito | Kosuke Nishida | Kyosuke Nishida | Itsumi Saito | J. Tomita | Kosuke Nishida
[1] Jiusheng Chen,et al. ProphetNet: Predicting Future N-gram for Sequence-to-Sequence Pre-training , 2020, EMNLP.
[2] Yen-Chun Chen,et al. Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting , 2018, ACL.
[3] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[4] Alexander M. Rush,et al. Bottom-Up Abstractive Summarization , 2018, EMNLP.
[5] Jianfeng Gao,et al. UniLMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training , 2020, ICML.
[6] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[7] Zita Marinho,et al. Jointly Extracting and Compressing Documents with Summary State Representations , 2019, NAACL.
[8] Philip Schlesinger,et al. Exploring the Limits: Europe’s Changing Communication Environment , 1997 .
[9] Mirella Lapata,et al. Don’t Give Me the Details, Just the Summary! Topic-Aware Convolutional Neural Networks for Extreme Summarization , 2018, EMNLP.
[10] Ian S. Dunn,et al. Exploring the Limits , 2009 .
[11] Min Sun,et al. A Unified Model for Extractive and Abstractive Summarization using Inconsistency Loss , 2018, ACL.
[12] Si Li,et al. Guiding Generation for Abstractive Text Summarization Based on Key Information Guide Network , 2018, NAACL.
[13] Ming Zhou,et al. HIBERT: Document Level Pre-training of Hierarchical Bidirectional Transformers for Document Summarization , 2019, ACL.
[14] Weijia Jia,et al. Improving Abstractive Document Summarization with Salient Information Modeling , 2019, ACL.
[15] Yang Liu,et al. Fine-tune BERT for Extractive Summarization , 2019, ArXiv.
[16] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[17] Yao Zhao,et al. PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization , 2020, ICML.
[18] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[19] Phil Blunsom,et al. Teaching Machines to Read and Comprehend , 2015, NIPS.
[20] Furu Wei,et al. Retrieve, Rerank and Rewrite: Soft Template Based Neural Summarization , 2018, ACL.
[21] Ming Zhou,et al. Selective Encoding for Abstractive Sentence Summarization , 2017, ACL.
[22] Xiaodong Liu,et al. Unified Language Model Pre-training for Natural Language Understanding and Generation , 2019, NeurIPS.
[23] Mirella Lapata,et al. Text Summarization with Pretrained Encoders , 2019, EMNLP.
[24] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[25] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[26] Xu Tan,et al. MASS: Masked Sequence to Sequence Pre-training for Language Generation , 2019, ICML.
[27] Wei Zhao,et al. Denoising based Sequence-to-Sequence Pre-training for Text Generation , 2019, EMNLP.