Synthesize, Prompt and Transfer: Zero-shot Conversational Question Generation with Pre-trained Language Model
暂无分享,去创建一个
[1] Bing Qin,et al. TAGNet: a tiny answer-guided network for conversational question generation , 2022, International Journal of Machine Learning and Cybernetics.
[2] Anima Anandkumar,et al. Test-Time Prompt Tuning for Zero-Shot Generalization in Vision-Language Models , 2022, NeurIPS.
[3] Graham Neubig,et al. Prompt Consistency for Zero-Shot Task Generalization , 2022, EMNLP.
[4] Hyung Won Chung,et al. What Language Model Architecture and Pretraining Objective Work Best for Zero-Shot Generalization? , 2022, ICML.
[5] Alona Fyshe,et al. Question Generation for Reading Comprehension Assessment by Modeling How and What to Ask , 2022, FINDINGS.
[6] Quoc V. Le,et al. Finetuned Language Models Are Zero-Shot Learners , 2021, ICLR.
[7] Quan Z. Sheng,et al. Conversational question answering: a survey , 2021, Knowledge and Information Systems.
[8] Jing Gu,et al. ChainCQG: Flow-Aware Conversational Question Generation , 2021, EACL.
[9] Jon Ander Campos,et al. DoQA - Accessing Domain-Specific FAQs via Conversational QA , 2020, ACL.
[10] Paul N. Bennett,et al. Generating Clarifying Questions for Information Retrieval , 2020, WWW.
[11] Paul N. Bennett,et al. Leading Conversational Search by Suggesting Useful Questions , 2020, WWW.
[12] Peter J. Liu,et al. PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization , 2019, ICML.
[13] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[14] Regina Barzilay,et al. Capturing Greater Context for Question Generation , 2019, AAAI.
[15] Piji Li,et al. Interconnected Question Generation with Coreference Alignment and Conversation Flow Modeling , 2019, ACL.
[16] Danqi Chen,et al. CoQA: A Conversational Question Answering Challenge , 2018, TACL.
[17] Eunsol Choi,et al. QuAC: Question Answering in Context , 2018, EMNLP.
[18] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[19] Harry Shum,et al. From Eliza to XiaoIce: challenges and opportunities with social chatbots , 2018, Frontiers of Information Technology & Electronic Engineering.
[20] Hannes Schulz,et al. Relevance of Unsupervised Metrics in Task-Oriented Dialogue for Evaluating Natural Language Generation , 2017, ArXiv.
[21] Xinya Du,et al. Learning to Ask: Neural Question Generation for Reading Comprehension , 2017, ACL.
[22] Ming Zhou,et al. Neural Question Generation from Text: A Preliminary Study , 2017, NLPCC.
[23] Christopher D. Manning,et al. Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.
[24] Philip Bachman,et al. NewsQA: A Machine Comprehension Dataset , 2016, Rep4NLP@ACL.
[25] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[26] Sanja Fidler,et al. Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[27] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[28] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[29] Wiebke Wagner,et al. Steven Bird, Ewan Klein and Edward Loper: Natural Language Processing with Python, Analyzing Text with the Natural Language Toolkit , 2010, Lang. Resour. Evaluation.
[30] Qiang Yang,et al. A Survey on Transfer Learning , 2010, IEEE Transactions on Knowledge and Data Engineering.
[31] Noah A. Smith,et al. Good Question! Statistical Ranking for Question Generation , 2010, NAACL.
[32] Fei-Fei Li,et al. ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[33] Xuanjing Huang,et al. CQG: A Simple and Effective Controlled Generation Framework for Multi-hop Question Generation , 2022, ACL.
[34] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[35] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[36] Yao Zhao,et al. Paragraph-level Neural Question Generation with Maxout Pointer and Gated Self-attention Networks , 2018, EMNLP.
[37] Noah A. Smith,et al. Automatic factual question generation from text , 2011 .