Self-Evaluation Guided Beam Search for Reasoning
暂无分享,去创建一个
Kenji Kawaguchi | Qizhe Xie | Junxian He | MingSung Kan | Xu Zhao | Yiran Zhao | Yuxi Xie
[1] Eric Michael Smith,et al. Llama 2: Open Foundation and Fine-Tuned Chat Models , 2023, ArXiv.
[2] B. Faltings,et al. REFINER: Reasoning Feedback on Intermediate Representations , 2023, ArXiv.
[3] Bodhisattwa Prasad Majumder,et al. Self-Refine: Iterative Refinement with Self-Feedback , 2023, NeurIPS.
[4] Naman Goyal,et al. LLaMA: Open and Efficient Foundation Language Models , 2023, ArXiv.
[5] Sida I. Wang,et al. Coder Reviewer Reranking for Code Generation , 2022, ICML.
[6] Geoffrey Irving,et al. Solving math word problems with process- and outcome-based feedback , 2022, ArXiv.
[7] William W. Cohen,et al. Program of Thoughts Prompting: Disentangling Computation from Reasoning for Numerical Reasoning Tasks , 2022, ArXiv.
[8] S. Gu,et al. Large Language Models Can Self-Improve , 2022, EMNLP.
[9] Song-Chun Zhu,et al. Dynamic Prompt Learning via Policy Gradient for Semi-structured Mathematical Reasoning , 2022, ICLR.
[10] Tom B. Brown,et al. Language Models (Mostly) Know What They Know , 2022, ArXiv.
[11] Yuhuai Wu,et al. Solving Quantitative Reasoning Problems with Language Models , 2022, NeurIPS.
[12] S. Gu,et al. Large Language Models are Zero-Shot Reasoners , 2022, NeurIPS.
[13] Haoming Jiang,et al. SeqZero: Few-shot Compositional Semantic Parsing with Sequential Prompts and Zero-shot Models , 2022, NAACL-HLT.
[14] Andrew M. Dai,et al. PaLM: Scaling Language Modeling with Pathways , 2022, J. Mach. Learn. Res..
[15] D. Schuurmans,et al. Self-Consistency Improves Chain of Thought Reasoning in Language Models , 2022, ICLR.
[16] Po-Sen Huang,et al. Scaling Language Models: Methods, Analysis & Insights from Training Gopher , 2021, ArXiv.
[17] David Bieber,et al. Show Your Work: Scratchpads for Intermediate Computation with Language Models , 2021, ArXiv.
[18] Mohammad Bavarian,et al. Training Verifiers to Solve Math Word Problems , 2021, ArXiv.
[19] Carrie J. Cai,et al. AI Chains: Transparent and Controllable Human-AI Interaction by Chaining Large Language Model Prompts , 2021, CHI.
[20] Tim Vieira,et al. Conditional Poisson Stochastic Beam Search , 2021, ArXiv.
[21] Wojciech Zaremba,et al. Evaluating Large Language Models Trained on Code , 2021, ArXiv.
[22] Navin Goyal,et al. Are NLP Models really able to Solve Simple Math Word Problems? , 2021, NAACL.
[23] Ana Marasović,et al. Teach Me to Explain: A Review of Datasets for Explainable Natural Language Processing , 2021, NeurIPS Datasets and Benchmarks.
[24] Jonathan Berant,et al. Did Aristotle Use a Laptop? A Question Answering Benchmark with Implicit Reasoning Strategies , 2021, Transactions of the Association for Computational Linguistics.
[25] Graham Neubig,et al. How Can We Know When Language Models Know? On the Calibration of Language Models for Question Answering , 2020, Transactions of the Association for Computational Linguistics.
[26] Noah A. Smith,et al. Measuring Association Between Labels and Free-Text Rationales , 2020, EMNLP.
[27] Chenyan Xiong,et al. Towards Interpretable Natural Language Understanding with Explanations as Latent Variables , 2020, NeurIPS.
[28] Ryan Cotterell,et al. Best-First Beam Search , 2020, Transactions of the Association for Computational Linguistics.
[29] Keh-Yih Su,et al. A Diverse Corpus for Evaluating and Developing English Math Word Problem Solvers , 2020, ACL.
[30] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[31] Bill Byrne,et al. On NMT Search Errors and Model Errors: Cat Got Your Tongue? , 2019, EMNLP.
[32] Yejin Choi,et al. The Curious Case of Neural Text Degeneration , 2019, ICLR.
[33] Max Welling,et al. Stochastic Beams and Where to Find Them: The Gumbel-Top-k Trick for Sampling Sequences Without Replacement , 2019, ICML.
[34] Joelle Pineau,et al. Language GANs Falling Short , 2018, ICLR.
[35] Yann Dauphin,et al. Hierarchical Neural Story Generation , 2018, ACL.
[36] Yong Yu,et al. Long Text Generation via Adversarial Training with Leaked Information , 2017, AAAI.
[37] Kilian Q. Weinberger,et al. On Calibration of Modern Neural Networks , 2017, ICML.
[38] Wang Ling,et al. Program Induction by Rationale Generation: Learning to Solve and Explain Algebraic Word Problems , 2017, ACL.
[39] Quoc V. Le,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[40] Alex Graves,et al. Sequence Transduction with Recurrent Neural Networks , 2012, ArXiv.
[41] Huan Sun,et al. Shepherd Pre-trained Language Models to Develop a Train of Thought: An Iterative Prompting Approach , 2022, ArXiv.
[42] Jonathan Berant,et al. CommonsenseQA: A Question Answering Challenge Targeting Commonsense Knowledge , 2019, NAACL.
[43] James H. Martin,et al. Speech and Language Processing: An Introduction to Natural Language Processing, Computational Linguistics, and Speech Recognition , 2000 .
[44] Geoffrey E. Hinton,et al. A Learning Algorithm for Boltzmann Machines , 1985, Cogn. Sci..