Towards Fewer Hallucinations in Knowledge-Grounded Dialogue Generation via Augmentative and Contrastive Knowledge-Dialogue

Existing knowledge-grounded open-domain dialogue generation models often face the hallucination problem, i.e. the dialogue generative model will persist in an inappropriate knowledge and generate responses that inconsistent with the facts. We argue that this problem mainly stems from the polarized optimization objectives and weak knowledge generation ability. To mitigate the hallucination, we take inspiration from human communicating that people will replay euphemistic responses for the unclear or unrecognizable knowledge, and propose an Augmentative and Contrastive Knowledge Dialogue Expansion Framework (ACK-DEF). ACK-DEF constructs the augmentative and contrastive knowledge dialogue samples, which consist of the knowledge of different degrees of errors and the response of manual design, to expand the original training set and smooth the polarized optimization objective that enables models to generate ground-truth with or without gold knowledge. Not only the knowledge, ACK-DEF also provides the tactful responses of manual design corresponding to the incomplete correct knowledge. Experimental results on the Wikipedia of Wizard dataset show that employing the ACK-DEF is effective to alleviate the hallucination problem.

[1]  M. de Rijke,et al.  Contrastive Learning Reduces Hallucination in Conversations , 2022, AAAI.

[2]  Haifeng Wang,et al.  PLATO-K: Internal and External Knowledge Enhanced Dialogue Generation , 2022, ArXiv.

[3]  Rui Yan,et al.  There Is No Standard Answer: Knowledge-Grounded Dialogue Generation with Adversarial Activated Multi-Reference Learning , 2022, EMNLP.

[4]  N. Zhang,et al.  Prompt Conditioned VAE: Enhancing Generative Replay for Lifelong Learning in Task-Oriented Dialogue , 2022, EMNLP.

[5]  David Vandyke,et al.  Prompting for a conversation: How to control a dialog model? , 2022, CAI.

[6]  Maosong Sun,et al.  Prompt Tuning for Discriminative Pre-trained Language Models , 2022, FINDINGS.

[7]  Mo Yu,et al.  On the Origin of Hallucinations in Conversational Models: Is it the Datasets or the Models? , 2022, NAACL.

[8]  M. Shoeybi,et al.  Multi-Stage Prompting for Knowledgeable Dialogue Generation , 2022, FINDINGS.

[9]  Jason Weston,et al.  Reason first, then respond: Modular Generation for Knowledge-infused Dialogue , 2021, EMNLP.

[10]  Jason Weston,et al.  Internet-Augmented Dialogue Generation , 2021, ACL.

[11]  David Reitter,et al.  Increasing Faithfulness in Knowledge-Grounded Dialogue with Controllable Features , 2021, ACL.

[12]  Shaoxiong Feng,et al.  Generating Relevant and Coherent Dialogue Responses using Self-Separated Conditional Variational AutoEncoders , 2021, ACL.

[13]  Jason Weston,et al.  Retrieval Augmentation Reduces Hallucination in Conversation , 2021, EMNLP.

[14]  Iryna Gurevych,et al.  The Curse of Dense Low-Dimensional Information Retrieval for Large Index Sizes , 2020, ACL.

[15]  Xiaofang Zhao,et al.  Group-wise Contrastive Learning for Neural Dialogue Generation , 2020, FINDINGS.

[16]  Dilek Z. Hakkani-Tür,et al.  Policy-Driven Neural Response Generation for Knowledge-Grounded Dialogue Systems , 2020, ArXiv.

[17]  Kaiming He,et al.  Improved Baselines with Momentum Contrastive Learning , 2020, ArXiv.

[18]  Dongyan Zhao,et al.  Low-Resource Knowledge-Grounded Dialogue Generation , 2020, ICLR.

[19]  Gunhee Kim,et al.  Sequential Latent Knowledge Selection for Knowledge-Grounded Dialogue , 2020, ICLR.

[20]  Geoffrey E. Hinton,et al.  A Simple Framework for Contrastive Learning of Visual Representations , 2020, ICML.

[21]  Hua Wu,et al.  PLATO: Pre-trained Dialogue Generation Model with Discrete Latent Variable , 2019, ACL.

[22]  Dilek Z. Hakkani-Tür,et al.  Topical-Chat: Towards Knowledge-Grounded Open-Domain Conversations , 2019, INTERSPEECH.

[23]  Xiyuan Zhang,et al.  Proactive Human-Machine Conversation with Explicit Conversation Goal , 2019, ACL.

[24]  Richard Csaky,et al.  Improving Neural Conversational Models with Entropy-Based Data Filtering , 2019, ACL.

[25]  Rongzhong Lian,et al.  Learning to Select Knowledge for Response Generation in Dialog Systems , 2019, IJCAI.

[26]  J. Weston,et al.  Wizard of Wikipedia: Knowledge-Powered Conversational agents , 2018, ICLR.

[27]  Alan W. Black,et al.  A Dataset for Document Grounded Conversations , 2018, EMNLP.

[28]  Verena Rieser,et al.  Better Conversations by Modeling, Filtering, and Optimizing for Coherence and Diversity , 2018, EMNLP.

[29]  Joelle Pineau,et al.  How NOT To Evaluate Your Dialogue System: An Empirical Study of Unsupervised Evaluation Metrics for Dialogue Response Generation , 2016, EMNLP.

[30]  Jianfeng Gao,et al.  A Diversity-Promoting Objective Function for Neural Conversation Models , 2015, NAACL.

[31]  Salim Roukos,et al.  Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.

[32]  George A. Miller,et al.  WordNet: A Lexical Database for English , 1995, HLT.

[33]  Natasa Milic-Frayling,et al.  Knowledge-Grounded Dialogue Generation with Term-level De-noising , 2021, FINDINGS.

[34]  Danqi Chen,et al.  of the Association for Computational Linguistics: , 2001 .