A Survey of the State-of-the-Art Models in Neural Abstractive Text Summarization
暂无分享,去创建一个
[1] Hang Li,et al. “ Tony ” DNN Embedding for “ Tony ” Selective Read for “ Tony ” ( a ) Attention-based Encoder-Decoder ( RNNSearch ) ( c ) State Update s 4 SourceVocabulary Softmax Prob , 2016 .
[2] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[3] Alexander M. Rush,et al. Abstractive Sentence Summarization with Attentive Recurrent Neural Networks , 2016, NAACL.
[4] Alon Lavie,et al. METEOR: An Automatic Metric for MT Evaluation with Improved Correlation with Human Judgments , 2005, IEEvaluation@ACL.
[5] Mor Naaman,et al. Newsroom: A Dataset of 1.3 Million Summaries with Diverse Extractive Strategies , 2018, NAACL.
[6] Marc'Aurelio Ranzato,et al. Sequence Level Training with Recurrent Neural Networks , 2015, ICLR.
[7] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[8] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[9] Jason Weston,et al. A Neural Attention Model for Abstractive Sentence Summarization , 2015, EMNLP.
[10] Sheikh Abujar,et al. Abstractive method of text summarization with sequence to sequence RNNs , 2019, 2019 10th International Conference on Computing, Communication and Networking Technologies (ICCCNT).
[11] Jeffrey L. Elman,et al. Finding Structure in Time , 1990, Cogn. Sci..
[12] Christopher D. Manning,et al. Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.
[13] Jiangtao Ren,et al. Summary-aware attention for social media short text abstractive summarization , 2020, Neurocomputing.
[14] Recent Trends in Deep Learning Based Abstractive Text Summarization , 2019, International journal of recent technology and engineering.
[15] Hans Peter Luhn,et al. The Automatic Creation of Literature Abstracts , 1958, IBM J. Res. Dev..
[16] Xiaojun Wan,et al. MultiSumm: Towards a Unified Model for Multi-Lingual Abstractive Summarization , 2020, AAAI.
[17] Xu Tan,et al. MASS: Masked Sequence to Sequence Pre-training for Language Generation , 2019, ICML.
[18] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[19] Martial Hebert,et al. Improving Multi-Step Prediction of Learned Time Series Models , 2015, AAAI.
[20] Krys J. Kochut,et al. Text Summarization Techniques: A Brief Survey , 2017, International Journal of Advanced Computer Science and Applications.
[22] Hiroyuki Shindo,et al. Length-controllable Abstractive Summarization by Guiding with Summary Prototype , 2020, ArXiv.
[23] Qingcai Chen,et al. LCSTS: A Large Scale Chinese Short Text Summarization Dataset , 2015, EMNLP.
[24] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[25] Yue Zhang,et al. Contrastive Attention Mechanism for Abstractive Sentence Summarization , 2019, EMNLP.
[26] Yen-Chun Chen,et al. Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting , 2018, ACL.
[27] William Yang Wang,et al. Deep Reinforcement Learning with Distributional Semantic Rewards for Abstractive Summarization , 2019, EMNLP.
[28] Suyanto Suyanto,et al. Indonesian Abstractive Text Summarization Using Bidirectional Gated Recurrent Unit , 2019, ICCSCI.
[29] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[30] Xipeng Qiu,et al. Pre-trained models for natural language processing: A survey , 2020, Science China Technological Sciences.
[31] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[32] Alexander M. Rush,et al. Bottom-Up Abstractive Summarization , 2018, EMNLP.
[33] Florian Matthes,et al. Automatic Text Summarization: A State-of-the-Art Review , 2020, ICEIS.
[34] Yoshua Bengio,et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.
[35] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[36] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[37] M. Maybury,et al. Automatic Summarization , 2002, Computational Linguistics.
[38] Jungang Xu,et al. A Survey on Neural Network Language Models , 2019, ArXiv.
[39] Bowen Zhou,et al. Abstractive Text Summarization using Sequence-to-sequence RNNs and Beyond , 2016, CoNLL.
[40] Richard T. Watson,et al. Analyzing the Past to Prepare for the Future: Writing a Literature Review , 2002, MIS Q..
[41] Yao Zhao,et al. PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization , 2020, ICML.
[42] Vincent Ng,et al. Abstractive Summarization: A Survey of the State of the Art , 2019, AAAI.
[43] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[44] Hazem M. Abbas,et al. Amharic Abstractive Text Summarization , 2020, ArXiv.
[45] Sebastian Ruder,et al. An overview of gradient descent optimization algorithms , 2016, Vestnik komp'iuternykh i informatsionnykh tekhnologii.
[46] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[47] Elena Lloret,et al. Text summarisation in progress: a literature review , 2011, Artificial Intelligence Review.
[48] Richard Socher,et al. A Deep Reinforced Model for Abstractive Summarization , 2017, ICLR.
[49] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[50] Dmitrii Aksenov,et al. Abstractive text summarization with neural sequence-to-sequence models , 2020 .
[51] Kang Yang,et al. Bidirectional Attentional Encoder-Decoder Model and Bidirectional Beam Search for Abstractive Summarization , 2018, ArXiv.
[52] Virapat Kieuvongngam,et al. Automatic Text Summarization of COVID-19 Medical Research Articles using BERT and GPT-2 , 2020, ArXiv.
[53] Zhen-Hua Ling,et al. Distraction-Based Neural Networks for Document Summarization , 2016, ArXiv.
[54] Nima Sanjabi. Abstractive text summarization with attention-based mechanism , 2018 .
[55] Haitao Huang,et al. Abstractive text summarization using LSTM-CNN based deep learning , 2018, Multimedia Tools and Applications.
[56] Dragomir R. Radev,et al. CO-Search: COVID-19 Information Retrieval with Semantic Search, Question Answering, and Abstractive Summarization , 2020, ArXiv.
[57] Xuedong Huang,et al. End-to-End Abstractive Summarization for Meetings , 2020, ArXiv.
[58] Kuldip K. Paliwal,et al. Bidirectional recurrent neural networks , 1997, IEEE Trans. Signal Process..
[59] Naren Ramakrishnan,et al. Neural Abstractive Text Summarization with Sequence-to-Sequence Models , 2018, Trans. Data Sci..