G ENERATE RATHER THAN R ETRIEVE : L ARGE L ANGU - AGE M ODELS ARE S TRONG C ONTEXT G ENERATORS
暂无分享,去创建一个
Wenhao Yu | Chenguang Zhu | Yichong Xu | Michael Zeng | Shuo Wang | Mingxuan Ju | Soumya Sanyal | Meng Jiang
[1] M. Shanahan,et al. Faithful Reasoning Using Large Language Models , 2022, ArXiv.
[2] Jane A. Yu,et al. Few-shot Learning with Retrieval Augmented Language Models , 2022, J. Mach. Learn. Res..
[3] Devendra Singh Sachan,et al. Questions Are All You Need to Train a Dense Passage Retriever , 2022, TACL.
[4] J. Dean,et al. Emergent Abilities of Large Language Models , 2022, Trans. Mach. Learn. Res..
[5] S. Gu,et al. Large Language Models are Zero-Shot Reasoners , 2022, NeurIPS.
[6] Colin Raffel,et al. Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning , 2022, NeurIPS.
[7] Xi Victoria Lin,et al. OPT: Open Pre-trained Transformer Language Models , 2022, ArXiv.
[8] Wen-tau Yih,et al. Autoregressive Search Engines: Generating Substrings as Document Identifiers , 2022, NeurIPS.
[9] S. Shalev-Shwartz,et al. Standing on the Shoulders of Giant Frozen Language Models , 2022, ArXiv.
[10] Andrew M. Dai,et al. PaLM: Scaling Language Modeling with Pathways , 2022, J. Mach. Learn. Res..
[11] Anastasia Chan,et al. GPT-3 and InstructGPT: technological dystopianism, utopianism, and “Contextual” perspectives in AI ethics and industry , 2022, AI and Ethics.
[12] Lisa Anne Hendricks,et al. Training Compute-Optimal Large Language Models , 2022, ArXiv.
[13] Angeliki Lazaridou,et al. Internet-augmented language models through few-shot prompting for open-domain question answering , 2022, ArXiv.
[14] Ryan J. Lowe,et al. Training language models to follow instructions with human feedback , 2022, NeurIPS.
[15] Alexander M. Rush,et al. PromptSource: An Integrated Development Environment and Repository for Natural Language Prompts , 2022, ACL.
[16] Dale Schuurmans,et al. Chain of Thought Prompting Elicits Reasoning in Large Language Models , 2022, NeurIPS.
[17] Edouard Grave,et al. Unsupervised Dense Information Retrieval with Contrastive Learning , 2021, Trans. Mach. Learn. Res..
[18] Quoc V. Le,et al. GLaM: Efficient Scaling of Language Models with Mixture-of-Experts , 2021, ICML.
[19] Po-Sen Huang,et al. Scaling Language Models: Methods, Analysis & Insights from Training Gopher , 2021, ArXiv.
[20] Shuohang Wang,et al. Leveraging Knowledge in Multilingual Commonsense Reasoning , 2021, FINDINGS.
[21] Ronan Le Bras,et al. Generated Knowledge Prompting for Commonsense Reasoning , 2021, ACL.
[22] Shuohang Wang,et al. KG-FiD: Infusing Knowledge Graph in Fusion-in-Decoder for Open-Domain Question Answering , 2021, ACL.
[23] Michael J.Q. Zhang,et al. SituatedQA: Incorporating Extra-Linguistic Contexts into QA , 2021, EMNLP.
[24] Quoc V. Le,et al. Finetuned Language Models Are Zero-Shot Learners , 2021, ICLR.
[25] Dani Yogatama,et al. End-to-End Training of Multi-Document Reader and Retriever for Open-Domain Question Answering , 2021, NeurIPS.
[26] Jordan L. Boyd-Graber,et al. Fool Me Twice: Entailment from Wikipedia Gamification , 2021, NAACL.
[27] Soujanya Poria,et al. Retrieving and Reading: A Comprehensive Survey on Open-domain Question Answering , 2021, ArXiv.
[28] Yelong Shen,et al. UnitedQA: A Hybrid Approach for Open Domain Question Answering , 2021, ACL.
[29] Edouard Grave,et al. Distilling Knowledge from Reader to Retriever for Question Answering , 2020, ICLR.
[30] Hua Wu,et al. RocketQA: An Optimized Training Approach to Dense Passage Retrieval for Open-Domain Question Answering , 2020, NAACL.
[31] Nicola De Cao,et al. Autoregressive Entity Retrieval , 2020, ICLR.
[32] Yelong Shen,et al. Generation-Augmented Retrieval for Open-Domain Question Answering , 2020, ACL.
[33] Nicola De Cao,et al. KILT: a Benchmark for Knowledge Intensive Language Tasks , 2020, NAACL.
[34] Edouard Grave,et al. Leveraging Passage Retrieval with Generative Models for Open Domain Question Answering , 2020, EACL.
[35] Christopher Potts,et al. Relevance-guided Supervision for OpenQA with ColBERT , 2020, Transactions of the Association for Computational Linguistics.
[36] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[37] Fabio Petroni,et al. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks , 2020, NeurIPS.
[38] Danqi Chen,et al. Dense Passage Retrieval for Open-Domain Question Answering , 2020, EMNLP.
[39] 知秀 柴田. 5分で分かる!? 有名論文ナナメ読み:Jacob Devlin et al. : BERT : Pre-training of Deep Bidirectional Transformers for Language Understanding , 2020 .
[40] Colin Raffel,et al. How Much Knowledge Can You Pack into the Parameters of a Language Model? , 2020, EMNLP.
[41] Alec Radford,et al. Scaling Laws for Neural Language Models , 2020, ArXiv.
[42] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[43] Sebastian Riedel,et al. Language Models as Knowledge Bases? , 2019, EMNLP.
[44] Ming-Wei Chang,et al. Natural Questions: A Benchmark for Question Answering Research , 2019, TACL.
[45] Ming-Wei Chang,et al. Latent Retrieval for Weakly Supervised Open Domain Question Answering , 2019, ACL.
[46] Yejin Choi,et al. The Curious Case of Neural Text Degeneration , 2019, ICLR.
[47] J. Weston,et al. Wizard of Wikipedia: Knowledge-Powered Conversational agents , 2018, ICLR.
[48] Andreas Vlachos,et al. FEVER: a Large-scale Dataset for Fact Extraction and VERification , 2018, NAACL.
[49] Eunsol Choi,et al. TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension , 2017, ACL.
[50] Jason Weston,et al. Reading Wikipedia to Answer Open-Domain Questions , 2017, ACL.
[51] Andrew Chou,et al. Semantic Parsing on Freebase from Question-Answer Pairs , 2013, EMNLP.
[52] Hugo Zaragoza,et al. The Probabilistic Relevance Framework: BM25 and Beyond , 2009, Found. Trends Inf. Retr..
[53] Weizhu Chen,et al. On the Advance of Making Language Models Better Reasoners , 2022, ArXiv.
[54] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.