Summarizing Chinese Medical Answer with Graph Convolution Networks and Question-focused Dual Attention

Online search engines are a popular source of medical information for users, where users can enter questions and obtain relevant answers. It is desirable to generate answer summaries for online search engines, particularly summaries that can reveal direct answers to questions. Moreover, answer summaries are expected to reveal the most relevant information in response to questions; hence, the summaries should be generated with a focus on the question, which is a challenging topic-focused summarization task. In this paper, we propose an approach that utilizes graph convolution networks and question-focused dual attention for Chinese medical answer summarization. We first organize the original long answer text into a medical concept graph with graph convolution networks to better understand the internal structure of the text and the correlation between medical concepts. Then, we introduce a question-focused dual attention mechanism to generate summaries relevant to questions. Experimental results demonstrate that the proposed model can generate more coherent and informative summaries compared with baseline models.

[1]  Xi Chen,et al.  Long-tail Relation Extraction via Knowledge Graph Embeddings and Graph Convolution Networks , 2019, NAACL.

[2]  Qianghuai Jia,et al.  Conceptualized Representation Learning for Chinese Biomedical Text Mining , 2020, ArXiv.

[3]  M. de Rijke,et al.  Summarizing Answers in Non-Factoid Community Question-Answering , 2017, WSDM.

[4]  Liang Zhou,et al.  Summarizing Answers for Complicated Questions , 2006, LREC.

[5]  Gholamreza Haffari,et al.  Graph-to-Sequence Learning using Gated Graph Neural Networks , 2018, ACL.

[6]  Christopher D. Manning,et al.  Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.

[7]  Yansong Feng,et al.  Graph2Seq: Graph to Sequence Learning with Attention-based Neural Networks , 2018, ArXiv.

[8]  Tiejun Zhao,et al.  Neural Document Summarization by Jointly Learning to Score and Select Sentences , 2018, ACL.

[9]  Hang Li,et al.  “ Tony ” DNN Embedding for “ Tony ” Selective Read for “ Tony ” ( a ) Attention-based Encoder-Decoder ( RNNSearch ) ( c ) State Update s 4 SourceVocabulary Softmax Prob , 2016 .

[10]  Wei Li,et al.  Coherent Comments Generation for Chinese Articles with a Graph-to-Sequence Model , 2019, ACL.

[11]  Mariana L. Neves,et al.  Entity-Supported Summarization of Biomedical Abstracts , 2016, BioTxtM@COLING 2016.

[12]  Yue Zhang,et al.  A Graph-to-Sequence Model for AMR-to-Text Generation , 2018, ACL.

[13]  Rada Mihalcea,et al.  TextRank: Bringing Order into Text , 2004, EMNLP.

[14]  A. Kogilavani,et al.  Ontology Enhanced Clustering Based Summarization of Medical Documents , 2009 .

[15]  Balaraman Ravindran,et al.  Diversity driven attention model for query-based abstractive summarization , 2017, ACL.

[16]  Jiebo Luo,et al.  Graph-based Neural Sentence Ordering , 2019, IJCAI.

[17]  Yiming Yang,et al.  XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.

[18]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[19]  Pablo Gervás,et al.  Concept-Graph Based Biomedical Automatic Summarization Using Ontologies , 2008, COLING 2008.

[20]  Alan W. Black,et al.  Boosting Dialog Response Generation , 2019, ACL.

[21]  Pengfei Liu,et al.  Heterogeneous Graph Neural Networks for Extractive Document Summarization , 2020, ACL.

[22]  Bowen Zhou,et al.  SummaRuNNer: A Recurrent Neural Network Based Sequence Model for Extractive Summarization of Documents , 2016, AAAI.

[23]  Jun Li,et al.  GraphSeq2Seq: Graph-Sequence-to-Sequence for Neural Machine Translation , 2018 .

[24]  Dietrich Klakow,et al.  Long-Span Language Models for Query-Focused Unsupervised Extractive Text Summarization , 2018, ECIR.

[25]  Wai Lam,et al.  Joint Learning of Answer Selection and Answer Summary Generation in Community Question Answering , 2019, AAAI.

[26]  Huajun Chen,et al.  Contrastive Triple Extraction with Generative Transformer , 2020, ArXiv.

[27]  Mirella Lapata,et al.  Neural Summarization by Extracting Sentences and Words , 2016, ACL.

[28]  Bowen Zhou,et al.  Abstractive Text Summarization using Sequence-to-sequence RNNs and Beyond , 2016, CoNLL.