Topic-Guided Abstractive Multi-Document Summarization

A critical point of multi-document summarization (MDS) is to learn the relations among various documents. In this paper, we propose a novel abstractive MDS model, in which we represent multiple documents as a heterogeneous graph, taking semantic nodes of different granularities into account, and then apply a graph-to-sequence framework to generate summaries. Moreover, we employ a neural topic model to jointly discover latent topics that can act as cross-document semantic units to bridge different documents and provide global information to guide the summary generation. Since topic extraction can be viewed as a special type of summarization that “summarizes” texts into a more abstract format, i.e., a topic distribution, we adopt a multi-task learning strategy to jointly train the topic and summarization module, allowing the promotion of each other. Experimental results on the Multi-News dataset demonstrate that our model outperforms previous state-of-the-art MDS models on both Rouge metrics and human evaluation, meanwhile learns high-quality topics.

[1]  Yue Wang,et al.  Topic-Aware Neural Keyphrase Generation for Social Media Language , 2019, ACL.

[2]  Pengfei Liu,et al.  Heterogeneous Graph Neural Networks for Extractive Document Summarization , 2020, ACL.

[3]  Lukasz Kaiser,et al.  Generating Wikipedia by Summarizing Long Sequences , 2018, ICLR.

[4]  Yang Wei Document summarization method based on heterogeneous graph , 2012, 2012 9th International Conference on Fuzzy Systems and Knowledge Discovery.

[5]  Jing Li,et al.  Topic Memory Networks for Short Text Classification , 2018, EMNLP.

[6]  Geoffrey E. Hinton,et al.  Layer Normalization , 2016, ArXiv.

[7]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[8]  Alexander M. Rush,et al.  Bottom-Up Abstractive Summarization , 2018, EMNLP.

[9]  Mirella Lapata,et al.  Neural Latent Extractive Document Summarization , 2018, EMNLP.

[10]  Bowen Zhou,et al.  SummaRuNNer: A Recurrent Neural Network Based Sequence Model for Extractive Summarization of Documents , 2016, AAAI.

[11]  Charles A. Sutton,et al.  Autoencoding Variational Inference For Topic Models , 2017, ICLR.

[12]  Rada Mihalcea,et al.  TextRank: Bringing Order into Text , 2004, EMNLP.

[13]  Pengfei Liu,et al.  Extractive Summarization as Text Matching , 2020, ACL.

[14]  Bo Chen,et al.  Friendly Topic Assistant for Transformer Based Abstractive Summarization , 2020, EMNLP.

[15]  Ming Zhou,et al.  HIBERT: Document Level Pre-training of Hierarchical Bidirectional Transformers for Document Summarization , 2019, ACL.

[16]  Vincent Ng,et al.  Abstractive Summarization: A Survey of the State of the Art , 2019, AAAI.

[17]  Dragomir R. Radev,et al.  Multi-News: A Large-Scale Multi-Document Summarization Dataset and Abstractive Hierarchical Model , 2019, ACL.

[18]  Xiaojun Wan,et al.  Multi-Granularity Interaction Network for Extractive and Abstractive Multi-Document Summarization , 2020, ACL.

[19]  Mirella Lapata,et al.  Hierarchical Transformers for Multi-Document Summarization , 2019, ACL.

[20]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[21]  Mirella Lapata,et al.  Don’t Give Me the Details, Just the Summary! Topic-Aware Convolutional Neural Networks for Extreme Summarization , 2018, EMNLP.

[22]  Yuanchao Liu,et al.  Enhancing Extractive Text Summarization with Topic-Aware Graph Neural Networks , 2020, COLING.

[23]  Phil Blunsom,et al.  Discovering Discrete Latent Topics with Neural Variational Inference , 2017, ICML.

[24]  Jackie Chi Kit Cheung,et al.  BanditSum: Extractive Summarization as a Contextual Bandit , 2018, EMNLP.

[25]  Michael Röder,et al.  Exploring the Space of Topic Coherence Measures , 2015, WSDM.

[26]  Pietro Liò,et al.  Graph Attention Networks , 2017, ICLR.

[27]  Franck Dernoncourt,et al.  A Discourse-Aware Attention Model for Abstractive Summarization of Long Documents , 2018, NAACL.

[28]  Christopher D. Manning,et al.  Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.

[29]  Michael I. Jordan,et al.  Latent Dirichlet Allocation , 2001, J. Mach. Learn. Res..

[30]  Xiaohui Yan,et al.  A biterm topic model for short texts , 2013, WWW.