Monotonic alignments for summarization

Abstract Summarization is the task that creates a summary with the major points of the original document. Deep learning plays an important role in both abstractive and extractive summary generations. While a number of models show that combining the two gives good results, this paper focuses on a pure abstractive method to generate good summaries. Our model is a stacked RNN network with a monotonic alignment mechanism. Monotonic alignment has an advantage because it produces the context that is in the same sequence as the original document, at the same time eliminating repeating sequences. To obtain monotonic alignment, this paper proposes two energies that are calculated using only the previous alignment state. We use sub-word method to reduce the rate of producing OOVs(Out of Vocabulary). The dropout is used for generalization and the residual connection to overcome gradient vanishing. We experiment on CNN/daily new and Reddits dataset. Our method out-performs the previous models with monotonic alignment by 4 ROUGE-1 points and achieves the results comparable to state of the art.

[1]  Yen-Chun Chen,et al.  Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting , 2018, ACL.

[2]  Yejin Choi,et al.  Deep Communicating Agents for Abstractive Summarization , 2018, NAACL.

[3]  Jie Tang,et al.  ArnetMiner: extraction and mining of academic social networks , 2008, KDD.

[4]  Christopher D. Manning,et al.  Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.

[5]  George Kurian,et al.  Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.

[6]  Bowen Zhou,et al.  Abstractive Text Summarization using Sequence-to-sequence RNNs and Beyond , 2016, CoNLL.

[7]  Franck Dernoncourt,et al.  A Discourse-Aware Attention Model for Abstractive Summarization of Long Documents , 2018, NAACL.

[8]  Yongbin Liu,et al.  Ensemble method to joint inference for knowledge extraction , 2017, Expert Syst. Appl..

[9]  Xu Sun,et al.  Global Encoding for Abstractive Summarization , 2018, ACL.

[10]  Piji Li,et al.  Deep Recurrent Generative Decoder for Abstractive Text Summarization , 2017, EMNLP.

[11]  Yongbin Liu,et al.  Empirical study on character level neural network classifier for Chinese text , 2019, Eng. Appl. Artif. Intell..

[12]  Colin Raffel,et al.  Monotonic Chunkwise Attention , 2017, ICLR.

[13]  Si Li,et al.  Guiding Generation for Abstractive Text Summarization Based on Key Information Guide Network , 2018, NAACL.

[14]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[15]  Angela Fan,et al.  Controllable Abstractive Summarization , 2017, NMT@ACL.

[16]  Rico Sennrich,et al.  Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.

[17]  Min Sun,et al.  A Unified Model for Extractive and Abstractive Summarization using Inconsistency Loss , 2018, ACL.

[18]  Chengjiang Li,et al.  XLORE2: Large-scale Cross-lingual Knowledge Graph Construction and Application , 2019, Data Intelligence.

[19]  Jing Zhang,et al.  AMiner: Search and Mining of Academic Social Networks , 2019, Data Intelligence.