Commonsense Knowledge Aware Conversation Generation with Graph Attention

Commonsense knowledge is vital to many natural language processing tasks. In this paper, we present a novel open-domain conversation generation model to demonstrate how large-scale commonsense knowledge can facilitate language understanding and generation. Given a user post, the model retrieves relevant knowledge graphs from a knowledge base and then encodes the graphs with a static graph attention mechanism, which augments the semantic information of the post and thus supports better understanding of the post. Then, during word generation, the model attentively reads the retrieved knowledge graphs and the knowledge triples within each graph to facilitate better generation through a dynamic graph attention mechanism. This is the first attempt that uses large-scale commonsense knowledge in conversation generation. Furthermore, unlike existing models that use knowledge triples (entities) separately and independently, our model treats each knowledge graph as a whole, which encodes more structured, connected semantic information in the graphs. Experiments show that the proposed model can generate more appropriate and informative responses than stateof-the-art baselines.

[1]  Marvin Minsky,et al.  Society of Mind: A Response to Four Reviews , 1991, Artif. Intell..

[2]  Jun Zhao,et al.  Conditional Generative Adversarial Networks for Commonsense Machine Comprehension , 2017, IJCAI.

[3]  Jason Weston,et al.  Translating Embeddings for Modeling Multi-relational Data , 2013, NIPS.

[4]  Hang Li,et al.  “ Tony ” DNN Embedding for “ Tony ” Selective Read for “ Tony ” ( a ) Attention-based Encoder-Decoder ( RNNSearch ) ( c ) State Update s 4 SourceVocabulary Softmax Prob , 2016 .

[5]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[6]  Yinong Long,et al.  A Knowledge Enhanced Generative Conversational Service Agent , 2017 .

[7]  Patrícia Cristina do Nascimento Souto,et al.  Creating Knowledge with and from the Differences: The Required Dialogicality and Dialogical Competences , 2015 .

[8]  Rui Yan,et al.  Sequence to Backward and Forward Sequences: A Content-Introducing Approach to Generative Short-Text Conversation , 2016, COLING.

[9]  Jianfeng Gao,et al.  A Neural Network Approach to Context-Sensitive Generation of Conversational Responses , 2015, NAACL.

[10]  Jianfeng Gao,et al.  A Diversity-Promoting Objective Function for Neural Conversation Models , 2015, NAACL.

[11]  Hongyu Lin,et al.  Reasoning with Heterogeneous Knowledge for Commonsense Machine Comprehension , 2017, EMNLP.

[12]  Wei-Ying Ma,et al.  Topic Aware Neural Response Generation , 2016, AAAI.

[13]  Sangdo Han,et al.  Exploiting knowledge base to generate responses for natural language dialog listening agents , 2015, SIGDIAL Conference.

[14]  Catherine Havasi,et al.  Representing General Relational Knowledge in ConceptNet 5 , 2012, LREC.

[15]  Per Linell,et al.  Dialogue in Focus Groups: Exploring Socially Shared Knowledge , 2007 .

[16]  Hang Li,et al.  Neural Responding Machine for Short-Text Conversation , 2015, ACL.

[17]  S. Eggins,et al.  Analysing Casual Conversation , 1996 .