A Unified Generative Retriever for Knowledge-Intensive Language Tasks via Prompt Learning
暂无分享,去创建一个
M. de Rijke | J. Guo | Xueqi Cheng | Yixing Fan | Y. Liu | Ruqing Zhang | Jiangui Chen
[1] Pascale Fung,et al. Survey of Hallucination in Natural Language Generation , 2022, ACM Comput. Surv..
[2] Hiroaki Hayashi,et al. Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing , 2021, ACM Comput. Surv..
[3] J. Guo,et al. CorpusBrain: Pre-train a Generative Retrieval Model for Knowledge-Intensive Language Tasks , 2022, CIKM.
[4] Qi Zhang,et al. A Neural Corpus Indexer for Document Retrieval , 2022, NeurIPS.
[5] Wen-tau Yih,et al. Autoregressive Search Engines: Generating Substrings as Document Identifiers , 2022, NeurIPS.
[6] J. Guo,et al. GERE: Generative Evidence Retrieval for Fact Verification , 2022, SIGIR.
[7] Liang Pang,et al. Match-Prompt: Improving Multi-task Generalization Ability for Neural Text Matching via Prompt Learning , 2022, CIKM.
[8] Ledell Yu Wu,et al. DynamicRetriever: A Pre-training Model-based IR System with Neither Sparse nor Dense Index , 2022, ArXiv.
[9] William W. Cohen,et al. Transformer Memory as a Differentiable Search Index , 2022, NeurIPS.
[10] Alexander M. Rush,et al. Multitask Prompted Training Enables Zero-Shot Task Generalization , 2021, ICLR.
[11] Minlie Huang,et al. PPT: Pre-trained Prompt Tuning for Few-shot Learning , 2021, ACL.
[12] Quoc V. Le,et al. Finetuned Language Models Are Zero-Shot Learners , 2021, ICLR.
[13] Jinho D. Choi,et al. The Stem Cell Hypothesis: Dilemma behind Multi-Task Learning with Transformer Encoders , 2021, EMNLP.
[14] Alfio Gliozzo,et al. Robust Retrieval Augmented Generation for Zero-shot Slot Filling , 2021, EMNLP.
[15] Zhengxiao Du,et al. GPT Understands, Too , 2021, AI Open.
[16] Fabio Petroni,et al. Multi-Task Retrieval for Knowledge-Intensive Tasks , 2021, ACL.
[17] Nicola De Cao,et al. Autoregressive Entity Retrieval , 2020, ICLR.
[18] Nicola De Cao,et al. KILT: a Benchmark for Knowledge Intensive Language Tasks , 2020, NAACL.
[19] Edouard Grave,et al. Leveraging Passage Retrieval with Generative Models for Open Domain Question Answering , 2020, EACL.
[20] Donald Metzler,et al. Rethinking Search: Making Domain Experts out of Dilettantes ∗ , 2021 .
[21] Luciano Floridi,et al. GPT-3: Its Nature, Scope, Limits, and Consequences , 2020, Minds and Machines.
[22] Fabio Petroni,et al. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks , 2020, NeurIPS.
[23] Danqi Chen,et al. Dense Passage Retrieval for Open-Domain Question Answering , 2020, EMNLP.
[24] Luke Zettlemoyer,et al. Zero-shot Entity Linking with Dense Entity Retrieval , 2019, ArXiv.
[25] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[26] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[27] Ming-Wei Chang,et al. Natural Questions: A Benchmark for Question Answering Research , 2019, TACL.
[28] Jason Weston,et al. ELI5: Long Form Question Answering , 2019, ACL.
[29] Roland Vollgraf,et al. FLAIR: An Easy-to-Use Framework for State-of-the-Art NLP , 2019, NAACL.
[30] Jonathan Berant,et al. MultiQA: An Empirical Investigation of Generalization and Transfer in Reading Comprehension , 2019, ACL.
[31] Jason Weston,et al. Wizard of Wikipedia: Knowledge-Powered Conversational agents , 2018, ICLR.
[32] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[33] Yoshua Bengio,et al. HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering , 2018, EMNLP.
[34] Christophe Gravier,et al. T-REx: A Large Scale Alignment of Natural Language with Knowledge Base Triples , 2018, LREC.
[35] Andreas Vlachos,et al. FEVER: a Large-scale Dataset for Fact Extraction and VERification , 2018, NAACL.
[36] Zhaochen Guo,et al. Robust named entity disambiguation with random walks , 2018, Semantic Web.
[37] Omer Levy,et al. Zero-Shot Relation Extraction via Reading Comprehension , 2017, CoNLL.
[38] Eunsol Choi,et al. TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension , 2017, ACL.
[39] Jason Weston,et al. Reading Wikipedia to Answer Open-Domain Questions , 2017, ACL.
[40] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[41] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[42] Gerhard Weikum,et al. Robust Disambiguation of Named Entities in Text , 2011, EMNLP.
[43] Hugo Zaragoza,et al. The Probabilistic Relevance Framework: BM25 and Beyond , 2009, Found. Trends Inf. Retr..
[44] Jason Weston,et al. A unified architecture for natural language processing: deep neural networks with multitask learning , 2008, ICML '08.
[45] Giovanni Manzini,et al. Opportunistic data structures with applications , 2000, Proceedings 41st Annual Symposium on Foundations of Computer Science.
[46] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[47] D. J. Wheeler,et al. A Block-sorting Lossless Data Compression Algorithm , 1994 .