KPT: Keyword-guided Pre-training for Grounded Dialog Generation

Incorporating external knowledge into the response generation process is essential to building more helpful and reliable dialog agents. However, collecting knowledge-grounded conversations is often costly, calling for a better pre-trained model for grounded dialog generation that generalizes well w.r.t. different types of knowledge. In this work, we propose KPT (Keyword-guided Pre-Training), a novel self-supervised pre-training method for grounded dialog generation without relying on extra knowledge annotation. Specifically, we use a pre-trained language model to extract the most uncertain tokens in the dialog as keywords. With these keywords, we construct two kinds of knowledge and pre-train a knowledge-grounded response generation model, aiming at handling two different scenarios: (1) the knowledge should be faithfully grounded; (2) it can be selectively used. For the former, the grounding knowledge consists of keywords extracted from the response. For the latter, the grounding knowledge is additionally augmented with keywords extracted from other utterances in the same dialog. Since the knowledge is extracted from the dialog itself, KPT can be easily performed on a large volume and variety of dialogue data. We considered three data sources (open-domain, task-oriented, conversational QA) with a total of 2.5M dialogues. We conduct extensive experiments on various few-shot knowledge-grounded generation tasks, including grounding on dialog acts, knowledge graphs, persona descriptions, and Wikipedia passages. Our comprehensive experiments and analyses demonstrate that KPT consistently outperforms state-of-the-art methods on these tasks with diverse grounding knowledge.

[1]  Carel van Niekerk,et al.  ConvLab-3: A Flexible Dialogue System Toolkit Based on a Unified Data Format , 2022, Conference on Empirical Methods in Natural Language Processing.

[2]  Fei Mi,et al.  Exploring Effective Information Utilization in Multi-Turn Topic-Driven Conversations , 2022, ArXiv.

[3]  Eric Michael Smith,et al.  BlenderBot 3: a deployed conversational agent that continually learns to responsibly engage , 2022, ArXiv.

[4]  Yinpei Dai,et al.  Unified Dialog Model Pre-training for Task-Oriented Dialog Understanding and Generation , 2022, SIGIR.

[5]  Bill Dolan,et al.  GODEL: Large-Scale Pre-Training for Goal-Directed Dialog , 2022, ArXiv.

[6]  Arun Tejasvi Chaganty,et al.  Dialog Inpainting: Turning Documents into Dialogs , 2022, ICML.

[7]  Yitong Li,et al.  PANGUBOT: Efficient Generative Dialogue Pre-training from Pre-trained Language Model , 2022, ArXiv.

[8]  Yitong Li,et al.  Pan More Gold from the Sand: Refining Open-domain Dialogue Training with Noisy Self-Retrieval Generation , 2022, COLING.

[9]  Renelito Delos Santos,et al.  LaMDA: Language Models for Dialog Applications , 2022, ArXiv.

[10]  Dragomir R. Radev,et al.  UnifiedSKG: Unifying and Multi-Tasking Structured Knowledge Grounding with Text-to-Text Language Models , 2022, EMNLP.

[11]  Baolin Peng,et al.  Soloist: Building Task Bots at Scale with Transfer Learning and Machine Teaching , 2021, Transactions of the Association for Computational Linguistics.

[12]  Hai Zhao,et al.  Dialogue-oriented Pre-training , 2021, FINDINGS.

[13]  Libo Qin,et al.  Language Model as an Annotator: Exploring DialoGPT for Dialogue Summarization , 2021, ACL.

[14]  Xiang Gao,et al.  RetGen: A Joint Framework for Retrieval and Grounded Text Generation Modeling , 2021, AAAI.

[15]  Jason Weston,et al.  Retrieval Augmentation Reduces Hallucination in Conversation , 2021, EMNLP.

[16]  Bill Byrne,et al.  TicketTalk: Toward human-level performance with end-to-end, transaction-based dialog systems , 2020, ACL.

[17]  Chunyan Miao,et al.  Keyword-Guided Neural Conversational Model , 2020, AAAI.

[18]  Kang Min Yoo,et al.  DialogBERT: Discourse-Aware Response Generation via Learning to Recover and Rank Utterances , 2020, AAAI.

[19]  Hannaneh Hajishirzi,et al.  UnifiedQA: Crossing Format Boundaries With a Single QA System , 2020, FINDINGS.

[20]  Mary Williamson,et al.  Recipes for Building an Open-Domain Chatbot , 2020, EACL.

[21]  Dongyan Zhao,et al.  Low-Resource Knowledge-Grounded Dialogue Generation , 2020, ICLR.

[22]  Jianfeng Gao,et al.  Results of the Multi-Domain Task-Completion Dialog Challenge , 2020, AAAI 2020.

[23]  Quoc V. Le,et al.  Towards a Human-like Open-Domain Chatbot , 2020, ArXiv.

[24]  Jianfeng Gao,et al.  DIALOGPT : Large-Scale Generative Pre-training for Conversational Response Generation , 2019, ACL.

[25]  Omer Levy,et al.  BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.

[26]  Colin Raffel,et al.  Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..

[27]  Raghav Gupta,et al.  Towards Scalable Multi-domain Conversational Agents: The Schema-Guided Dialogue Dataset , 2019, AAAI.

[28]  Bill Byrne,et al.  Taskmaster-1: Toward a Realistic and Diverse Dialog Dataset , 2019, EMNLP.

[29]  Anuj Kumar Goyal,et al.  MultiWOZ 2.1: A Consolidated Multi-Domain Dialogue Dataset with State Corrections and State Tracking Baselines , 2019, LREC.

[30]  Seungwhan Moon,et al.  OpenDialKG: Explainable Conversational Reasoning with Attention-based Walks over Knowledge Graphs , 2019, ACL.

[31]  Tiancheng Zhao,et al.  Pretraining Methods for Dialog Context Representation Learning , 2019, ACL.

[32]  Jianfeng Gao,et al.  Multi-Domain Task-Completion Dialog Challenge , 2019 .

[33]  Jason Weston,et al.  Wizard of Wikipedia: Knowledge-Powered Conversational agents , 2018, ICLR.

[34]  Stefan Ultes,et al.  MultiWOZ - A Large-Scale Multi-Domain Wizard-of-Oz Dataset for Task-Oriented Dialogue Modelling , 2018, EMNLP.

[35]  Xiaoyan Zhu,et al.  Commonsense Knowledge Aware Conversation Generation with Graph Attention , 2018, IJCAI.

[36]  Dongyan Zhao,et al.  An Ensemble of Retrieval-Based and Generation-Based Human-Computer Conversation Systems , 2018, IJCAI.

[37]  Matt Post,et al.  A Call for Clarity in Reporting BLEU Scores , 2018, WMT.

[38]  Jason Weston,et al.  Personalizing Dialogue Agents: I have a dog, do you have pets too? , 2018, ACL.

[39]  Xiaoyu Shen,et al.  DailyDialog: A Manually Labelled Multi-turn Dialogue Dataset , 2017, IJCNLP.

[40]  Jason Weston,et al.  ParlAI: A Dialog Research Software Platform , 2017, EMNLP.

[41]  Christopher D. Manning,et al.  Key-Value Retrieval Networks for Task-Oriented Dialogue , 2017, SIGDIAL Conference.

[42]  Ming-Wei Chang,et al.  A Knowledge-Grounded Neural Conversation Model , 2017, AAAI.

[43]  Jianfeng Gao,et al.  A Human Generated MAchine Reading COmprehension Dataset , 2018 .

[44]  Chin-Yew Lin,et al.  ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.

[45]  Salim Roukos,et al.  Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.

[46]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[47]  Bill Dolan,et al.  Grounded Response Generation Task at DSTC7 , 2019 .

[48]  Ilya Sutskever,et al.  Language Models are Unsupervised Multitask Learners , 2019 .

[49]  Danqi Chen,et al.  of the Association for Computational Linguistics: , 2001 .

[50]  Very Large Corpora Empirical Methods in Natural Language Processing , 1999 .