暂无分享,去创建一个
Gökhan Tür | Dilek Z. Hakkani-Tür | Alexandros Papangelis | Mahdi Namazifar | Gökhan Tür | Mahdi Namazifar | A. Papangelis
[1] Hui Xiong,et al. A Comprehensive Survey on Transfer Learning , 2019, Proceedings of the IEEE.
[2] Richard Socher,et al. The Natural Language Decathlon: Multitask Learning as Question Answering , 2018, ArXiv.
[3] Gökhan Tür,et al. Towards Zero-Shot Frame Semantic Parsing for Domain Scaling , 2017, INTERSPEECH.
[4] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[5] Dilek Z. Hakkani-Tür,et al. Robust Zero-Shot Cross-Domain Slot Filling with Example Values , 2019, ACL.
[6] Yi Zhang,et al. Learning to Classify Intents and Slot Labels Given a Handful of Examples , 2020, NLP4CONVAI.
[7] Philip S. Yu,et al. Joint Slot Filling and Intent Detection via Capsule Neural Networks , 2018, ACL.
[8] Irfan Ahmad,et al. A Survey on Transfer Learning in Natural Language Processing , 2020, ArXiv.
[9] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[10] Kevin Gimpel,et al. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations , 2019, ICLR.
[11] Dilek Z. Hakkani-Tür,et al. Dialog State Tracking: A Neural Reading Comprehension Approach , 2019, SIGdial.
[12] Percy Liang,et al. Know What You Don’t Know: Unanswerable Questions for SQuAD , 2018, ACL.
[13] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[14] Thomas Wolf,et al. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter , 2019, ArXiv.
[15] Thomas Wolf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[16] George R. Doddington,et al. The ATIS Spoken Language Systems Pilot Corpus , 1990, HLT.
[17] Dilek Z. Hakkani-Tür,et al. From Machine Reading Comprehension to Dialogue State Tracking: Bridging the Gap , 2020, NLP4CONVAI.
[18] Gökhan Tür,et al. What is left to be understood in ATIS? , 2010, 2010 IEEE Spoken Language Technology Workshop.
[19] Uthayasanker Thayasivam,et al. Meta Learning for Few-Shot Joint Intent Detection and Slot-Filling , 2020, ICML 2020.
[20] Andrea Madotto,et al. Language Models as Few-Shot Learner for Task-Oriented Dialogue Systems , 2020, ArXiv.
[21] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[22] Matthew Henderson,et al. Span-ConveRT: Few-shot Span Extraction for Dialog with Pretrained Conversational Representations , 2020, ACL.
[23] Wen Wang,et al. BERT for Joint Intent Classification and Slot Filling , 2019, ArXiv.
[24] Chao Yang,et al. A Survey on Deep Transfer Learning , 2018, ICANN.
[25] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.