A Dual Prompt Learning Framework for Few-Shot Dialogue State Tracking

Dialogue State Tracking (DST) module is an essential component of task-oriented dialog systems to understand users’ goals and needs. Collecting dialogue state labels including slots and values can be costly, requiring experts to annotate all (slot, value) information for each turn in dialogues. It is also difficult to define all possible slots and values in advance, especially with the wide application of dialogue systems in more and more new-rising applications. In this paper, we focus on improving DST module to generate dialogue states in circumstances with limited annotations and knowledge about slot ontology. To this end, we design a dual prompt learning framework for few-shot DST. The dual framework aims to explore how to utilize the language understanding and generation capabilities of pre-trained language models for DST efficiently. Specifically, we consider the learning of slot generation and value generation as dual tasks, and two kinds of prompts are designed based on this dual structure to incorporate task-related knowledge of these two tasks respectively. In this way, the DST task can be formulated as a language modeling task efficiently under few-shot settings. To evaluate the proposed framework, we conduct experiments on two task-oriented dialogue datasets. The results demonstrate that the proposed method not only outperforms existing state-of-the-art few-shot methods, but also can generate unseen slots. It indicates that DST-related knowledge can be probed from pre-trained language models and utilized to address low-resource DST efficiently with the help of prompt learning.

[1]  Elman Mansimov,et al.  Multi-Task Pre-Training for Plug-and-Play Task-Oriented Dialogue System , 2021, ACL.

[2]  Mari Ostendorf,et al.  Dialogue State Tracking with a Language Model using Schema-Driven Prompting , 2021, EMNLP.

[3]  Zhou Yu,et al.  Zero-Shot Dialogue State Tracking via Cross-Task Transfer , 2021, EMNLP.

[4]  Baolin Peng,et al.  Soloist: Building Task Bots at Scale with Transfer Learning and Machine Teaching , 2021, Transactions of the Association for Computational Linguistics.

[5]  Hiroaki Hayashi,et al.  Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing , 2021, ACM Comput. Surv..

[6]  Leyang Cui,et al.  Template-Based Named Entity Recognition Using BART , 2021, FINDINGS.

[7]  Zhou Yu,et al.  Leveraging Slot Descriptions for Zero-Shot Cross-Domain Dialogue StateTracking , 2021, NAACL.

[8]  Bill Byrne,et al.  Knowledge-Aware Graph-Enhanced GPT-2 for Dialogue State Tracking , 2021, EMNLP.

[9]  Zhengxiao Du,et al.  GPT Understands, Too , 2021, AI Open.

[10]  Alexander M. Rush,et al.  How many data points is a prompt worth? , 2021, NAACL.

[11]  Shang-Wen Li,et al.  Zero-shot Generalization in Dialog State Tracking through Generative Question Answering , 2021, EACL.

[12]  Dilek Z. Hakkani-Tür,et al.  Few Shot Dialogue State Tracking using Meta-learning , 2021, EACL.

[13]  Danqi Chen,et al.  Making Pre-trained Language Models Better Few-shot Learners , 2021, ACL.

[14]  Wanxiang Che,et al.  C2C-GenDA: Cluster-to-Cluster Generation for Data Augmentation of Slot Filling , 2020, AAAI.

[15]  Caiming Xiong,et al.  Improving Limited Labeled Dialogue State Tracking with Self-Supervision , 2020, FINDINGS.

[16]  Pascale Fung,et al.  MinTL: Minimalist Transfer Learning for Task-Oriented Dialogue Systems , 2020, EMNLP.

[17]  Qingkai Min,et al.  Dialogue State Induction Using Neural Latent Variable Models , 2020, IJCAI.

[18]  Mark Chen,et al.  Language Models are Few-Shot Learners , 2020, NeurIPS.

[19]  R. Socher,et al.  A Simple Language Model for Task-Oriented Dialogue , 2020, Neural Information Processing Systems.

[20]  Monica S. Lam,et al.  Zero-Shot Transfer Learning with Synthesized Data for Multi-Domain Dialogue State Tracking , 2020, ACL.

[21]  Richard Socher,et al.  TOD-BERT: Pre-trained Natural Language Understanding for Task-Oriented Dialogue , 2020, EMNLP.

[22]  Dilek Z. Hakkani-Tür,et al.  From Machine Reading Comprehension to Dialogue State Tracking: Bridging the Gap , 2020, NLP4CONVAI.

[23]  Timo Schick,et al.  Exploiting Cloze-Questions for Few-Shot Text Classification and Natural Language Inference , 2020, EACL.

[24]  Frank F. Xu,et al.  How Can We Know What Language Models Know? , 2019, Transactions of the Association for Computational Linguistics.

[25]  Li Zhou,et al.  Multi-domain Dialogue State Tracking as Dynamic Knowledge Graph Enhanced Question Answering , 2019, ArXiv.

[26]  Raghav Gupta,et al.  Towards Scalable Multi-domain Conversational Agents: The Schema-Guided Dialogue Dataset , 2019, AAAI.

[27]  Bill Byrne,et al.  Taskmaster-1: Toward a Realistic and Diverse Dialog Dataset , 2019, EMNLP.

[28]  Alexander M. Rush,et al.  Commonsense Knowledge Mining from Pretrained Models , 2019, EMNLP.

[29]  Dilek Z. Hakkani-Tür,et al.  Dialog State Tracking: A Neural Reading Comprehension Approach , 2019, SIGdial.

[30]  Richard Socher,et al.  Transferable Multi-Domain State Generator for Task-Oriented Dialogue Systems , 2019, ACL.

[31]  Stefan Ultes,et al.  MultiWOZ - A Large-Scale Multi-Domain Wizard-of-Oz Dataset for Task-Oriented Dialogue Modelling , 2018, EMNLP.

[32]  Sungjin Lee,et al.  Zero-Shot Adaptive Transfer for Conversational Language Understanding , 2018, AAAI.

[33]  Min-Yen Kan,et al.  Sequicity: Simplifying Task-oriented Dialogue Systems with Single Sequence-to-Sequence Architectures , 2018, ACL.

[34]  Vaibhava Goel,et al.  Self-Critical Sequence Training for Image Captioning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[35]  David Vandyke,et al.  A Network-based End-to-End Trainable Task-oriented Dialogue System , 2016, EACL.

[36]  Alexander I. Rudnicky,et al.  Unsupervised induction and filling of semantic slots for spoken dialogue systems using frame-semantic parsing , 2013, 2013 IEEE Workshop on Automatic Speech Recognition and Understanding.

[37]  Milica Gasic,et al.  POMDP-Based Statistical Spoken Dialog Systems: A Review , 2013, Proceedings of the IEEE.

[38]  C. Fillmore FRAME SEMANTICS AND THE NATURE OF LANGUAGE * , 1976 .

[39]  Zhou Yu,et al.  Discovering Dialogue Slots with Weak Supervision , 2021, ACL.

[40]  Percy Liang,et al.  Prefix-Tuning: Optimizing Continuous Prompts for Generation , 2021, ACL.

[41]  Roi Reichart,et al.  PADA: A Prompt-based Autoregressive Approach for Adaptation to Unseen Domains , 2021, ArXiv.

[42]  Dilek Z. Hakkani-Tür,et al.  MultiWOZ 2.1: Multi-Domain Dialogue State Corrections and State Tracking Baselines , 2019, ArXiv.

[43]  Ilya Sutskever,et al.  Language Models are Unsupervised Multitask Learners , 2019 .