Generating Abstractive Summaries with Finetuned Language Models
暂无分享,去创建一个
[1] Benno Stein,et al. TIRA Integrated Research Architecture , 2019, Information Retrieval Evaluation in a Changing World.
[2] Yen-Chun Chen,et al. Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting , 2018, ACL.
[3] Christopher D. Manning,et al. Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.
[4] Alexandros Potamianos,et al. An Embarrassingly Simple Approach for Transfer Learning from Pretrained Language Models , 2019, NAACL.
[5] Phil Blunsom,et al. Teaching Machines to Read and Comprehend , 2015, NIPS.
[6] Richard Socher,et al. A Deep Reinforced Model for Abstractive Summarization , 2017, ICLR.
[7] Ji Wang,et al. Pretraining-Based Natural Language Generation for Text Summarization , 2019, CoNLL.
[8] Alexander M. Rush,et al. Encoder-Agnostic Adaptation for Conditional Language Generation , 2019, ArXiv.
[9] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[10] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[11] Piji Li,et al. Actor-Critic based Training Framework for Abstractive Summarization , 2018, ArXiv.
[12] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[13] Benno Stein,et al. TL;DR: Mining Reddit to Learn Automatic Summarization , 2017, NFiS@EMNLP.
[14] Christopher D. Manning,et al. Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.
[15] Jason Weston,et al. A Neural Attention Model for Abstractive Sentence Summarization , 2015, EMNLP.
[16] Gunhee Kim,et al. Abstractive Summarization of Reddit Posts with Multi-level Memory Networks , 2018, NAACL.
[17] Ido Dagan,et al. How to Compare Summarizers without Target Length? Pitfalls, Solutions and Re-Examination of the Neural Summarization Literature , 2019, Proceedings of the Workshop on Methods for Optimizing and Evaluating Neural Language Generation.
[18] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[19] Hang Li,et al. “ Tony ” DNN Embedding for “ Tony ” Selective Read for “ Tony ” ( a ) Attention-based Encoder-Decoder ( RNNSearch ) ( c ) State Update s 4 SourceVocabulary Softmax Prob , 2016 .
[20] Navdeep Jaitly,et al. Pointer Networks , 2015, NIPS.
[21] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[22] Mona Attariyan,et al. Parameter-Efficient Transfer Learning for NLP , 2019, ICML.
[23] Bowen Zhou,et al. Abstractive Text Summarization using Sequence-to-sequence RNNs and Beyond , 2016, CoNLL.
[24] Alexander M. Rush,et al. Bottom-Up Abstractive Summarization , 2018, EMNLP.