Inner Monologue: Embodied Reasoning through Planning with Language Models
暂无分享,去创建一个
Peter R. Florence | S. Levine | Pierre Sermanet | Karol Hausman | Andy Zeng | Yevgen Chebotar | Igor Mordatch | Ted Xiao | Jonathan Tompson | Brian Ichter | F. Xia | Harris Chan | Jacky Liang | Wenlong Huang | Jacky Liang | Noah Brown | Tomas Jackson | Linda Luu | P. Sermanet
[1] Ian S. Fischer,et al. Deep Hierarchical Planning from Pixels , 2022, NeurIPS.
[2] Petko Georgiev,et al. Intra-agent speech permits zero-shot task acquisition , 2022, NeurIPS.
[3] Pierre-Yves Oudeyer,et al. Vygotskian Autotelic Artificial Intelligence: Language and Culture Internalization for Human-Like AI , 2022, ArXiv.
[4] S. Gu,et al. Large Language Models are Zero-Shot Reasoners , 2022, ArXiv.
[5] Oriol Vinyals,et al. Flamingo: a Visual Language Model for Few-Shot Learning , 2022, ArXiv.
[6] Abhinav Gupta,et al. Can Foundation Models Perform Zero-Shot Task Specification For Robot Manipulation? , 2022, L4DC.
[7] Oier Mees,et al. What Matters in Language Conditioned Robotic Imitation Learning Over Unstructured Data , 2022, IEEE Robotics and Automation Letters.
[8] Can language models learn from explanations in context? , 2022, ArXiv.
[9] Andrew M. Dai,et al. PaLM: Scaling Language Modeling with Pathways , 2022, J. Mach. Learn. Res..
[10] S. Levine,et al. Do As I Can, Not As I Say: Grounding Language in Robotic Affordances , 2022, CoRL.
[11] Adrian S. Wong,et al. Socratic Models: Composing Zero-Shot Multimodal Reasoning with Language , 2022, ICLR.
[12] J. Tenenbaum,et al. Inventing Relational State and Action Abstractions for Effective and Efficient Bilevel Planning , 2022, ArXiv.
[13] Ryan J. Lowe,et al. Training language models to follow instructions with human feedback , 2022, NeurIPS.
[14] A. Torralba,et al. Pre-Trained Language Models for Interactive Decision-Making , 2022, NeurIPS.
[15] Dale Schuurmans,et al. Chain of Thought Prompting Elicits Reasoning in Large Language Models , 2022, ArXiv.
[16] P. Abbeel,et al. Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied Agents , 2022, ICML.
[17] S. Levine,et al. Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning , 2021, ICLR.
[18] A. Torralba,et al. Skill Induction and Planning with Latent Language , 2021, ACL.
[19] Toki Migimatsu,et al. Grounding Predicates through Actions , 2021, 2022 International Conference on Robotics and Automation (ICRA).
[20] Bolei Zhou,et al. PlaTe: Visually-Grounded Planning With Transformers in Procedural Tasks , 2021, IEEE Robotics and Automation Letters.
[21] Quoc V. Le,et al. Finetuned Language Models Are Zero-Shot Learners , 2021, ICLR.
[22] Adams Wei Yu,et al. SimVLM: Simple Visual Language Model Pretraining with Weak Supervision , 2021, ICLR.
[23] Yin Cui,et al. Open-vocabulary Object Detection via Vision and Language Knowledge Distillation , 2021, ICLR.
[24] David Bieber,et al. Show Your Work: Scratchpads for Intermediate Computation with Language Models , 2021, ArXiv.
[25] R. Mottaghi,et al. Simple but Effective: CLIP Embeddings for Embodied AI , 2021, 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[26] Dieter Fox,et al. CLIPort: What and Where Pathways for Robotic Manipulation , 2021, CoRL.
[27] Jason Baldridge,et al. MURAL: Multimodal, Multitask Retrieval Across Languages , 2021, ArXiv.
[28] Alessandro Suglia,et al. Embodied BERT: A Transformer Model for Embodied, Language-guided Visual Task Completion , 2021, ArXiv.
[29] Wojciech Zaremba,et al. Evaluating Large Language Models Trained on Code , 2021, ArXiv.
[30] Yann LeCun,et al. MDETR - Modulated Detection for End-to-End Multi-Modal Understanding , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[31] S. Levine,et al. MT-Opt: Continuous Multi-Task Robotic Reinforcement Learning at Scale , 2021, ArXiv.
[32] Patricio A. Vela,et al. A Joint Network for Grasp Detection Conditioned on Natural Language Commands , 2021, 2021 IEEE International Conference on Robotics and Automation (ICRA).
[33] Dorsa Sadigh,et al. ELLA: Exploration through Learned Language Abstraction , 2021, NeurIPS.
[34] Ilya Sutskever,et al. Learning Transferable Visual Models From Natural Language Supervision , 2021, ICML.
[35] Silvio Savarese,et al. ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for Mobile Manipulation , 2020, ArXiv.
[36] Corey Lynch,et al. Language Conditioned Imitation Learning Over Unstructured Data , 2020, Robotics: Science and Systems.
[37] Pierre-Yves Oudeyer,et al. Grounding Language to Autonomously-Acquired Skills via Goal Generation , 2021, ICLR.
[38] Ross A. Knepper,et al. Few-shot Object Grounding and Mapping for Natural Language Robot Instruction Following , 2020, CoRL.
[39] Peter R. Florence,et al. Transporter Networks: Rearranging the Visual World for Robotic Manipulation , 2020, CoRL.
[40] Chitta Baral,et al. Language-Conditioned Imitation Learning for Robot Manipulation Tasks , 2020, NeurIPS.
[41] Peter Alexander Jansen. Visually-Grounded Planning without Vision: Language Models Infer Detailed Plans from High-level Instructions , 2020, FINDINGS.
[42] Geoffrey E. Hinton,et al. Big Self-Supervised Models are Strong Semi-Supervised Learners , 2020, NeurIPS.
[43] Karol Hausman,et al. Modeling Long-horizon Tasks as Sequential Interaction Landscapes , 2020, CoRL.
[44] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[45] Hong-Yuan Mark Liao,et al. YOLOv4: Optimal Speed and Accuracy of Object Detection , 2020, ArXiv.
[46] Karol Hausman,et al. Thinking While Moving: Deep Reinforcement Learning with Concurrent Control , 2020, ICLR.
[47] Colin Raffel,et al. How Much Knowledge Can You Pack into the Parameters of a Language Model? , 2020, EMNLP.
[48] Yoav Goldberg,et al. oLMpics-On What Language Model Pre-training Captures , 2019, Transactions of the Association for Computational Linguistics.
[49] Frank F. Xu,et al. How Can We Know What Language Models Know? , 2019, Transactions of the Association for Computational Linguistics.
[50] Jason J. Corso,et al. Unified Vision-Language Pre-Training for Image Captioning and VQA , 2019, AAAI.
[51] Chelsea Finn,et al. Hierarchical Foresight: Self-Supervised Learning of Long-Horizon Tasks via Visual Subgoal Generation , 2019, ICLR.
[52] Chuang Gan,et al. Once for All: Train One Network and Specialize it for Efficient Deployment , 2019, ICLR.
[53] Silvio Savarese,et al. HRL4IN: Hierarchical Reinforcement Learning for Interactive Navigation with Mobile Manipulators , 2019, CoRL.
[54] Ross A. Knepper,et al. Learning to Map Natural Language Instructions to Physical Quadcopter Control using Simulated Flight , 2019, CoRL.
[55] Silvio Savarese,et al. Regression Planning Networks , 2019, NeurIPS.
[56] Sebastian Riedel,et al. Language Models as Knowledge Bases? , 2019, EMNLP.
[57] Alexander M. Rush,et al. Commonsense Knowledge Mining from Pretrained Models , 2019, EMNLP.
[58] Iryna Gurevych,et al. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks , 2019, EMNLP.
[59] Stefan Lee,et al. ViLBERT: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks , 2019, NeurIPS.
[60] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[61] Chelsea Finn,et al. Language as an Abstraction for Hierarchical Deep Reinforcement Learning , 2019, NeurIPS.
[62] Sergey Levine,et al. Search on the Replay Buffer: Bridging Planning and Reinforcement Learning , 2019, NeurIPS.
[63] Jieping Ye,et al. Object Detection in 20 Years: A Survey , 2019, Proceedings of the IEEE.
[64] Dieter Fox,et al. Prospection: Interpretable plans from language by predicting the future , 2019, 2019 International Conference on Robotics and Automation (ICRA).
[65] Fadime Sener,et al. Zero-Shot Anticipation for Instructional Activities , 2018, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[66] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[67] Pieter Abbeel,et al. Learning Plannable Representations with Causal InfoGAN , 2018, NeurIPS.
[68] Patricio A. Vela,et al. Real-World Multiobject, Multigrasp Detection , 2018, IEEE Robotics and Automation Letters.
[69] Allan Jabri,et al. Universal Planning Networks: Learning Generalizable Representations for Visuomotor Control , 2018, ICML.
[70] Marc Toussaint,et al. Differentiable Physics and Stable Modes for Tool-Use and Manipulation Planning , 2018, Robotics: Science and Systems.
[71] Silvio Savarese,et al. Neural Task Programming: Learning to Generalize Across Hierarchical Tasks , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[72] Marc Toussaint,et al. Logic-Geometric Programming: An Optimization-Based Approach to Combined Task and Motion Planning , 2015, IJCAI.
[73] Xiaolin Hu,et al. Recurrent convolutional neural network for object recognition , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[74] Kaiming He,et al. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[75] Margaret Mitchell,et al. VQA: Visual Question Answering , 2015, International Journal of Computer Vision.
[76] Pieter Abbeel,et al. Combined task and motion planning through an extensible planner-independent interface layer , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).
[77] Ross A. Knepper,et al. Asking for Help Using Inverse Semantics , 2014, Robotics: Science and Systems.
[78] Leslie Pack Kaelbling,et al. Integrated task and motion planning in belief space , 2013, Int. J. Robotics Res..
[79] Honglak Lee,et al. Deep learning for detecting robotic grasps , 2013, Int. J. Robotics Res..
[80] Stefanie Tellex,et al. Interpreting and Executing Recipes with a Cooking Robot , 2012, ISER.
[81] Matthew R. Walter,et al. Understanding Natural Language Commands for Robotic Navigation and Mobile Manipulation , 2011, AAAI.
[82] Leslie Pack Kaelbling,et al. Hierarchical Planning in the Now , 2010, Bridging the Gap Between Task and Motion Planning.
[83] Stefanie Tellex,et al. Toward understanding natural language directions , 2010, 2010 5th ACM/IEEE International Conference on Human-Robot Interaction (HRI).
[84] Stefanie Tellex,et al. Grounding Verbs of Motion in Natural Language Commands to Robots , 2010, ISER.
[85] L. Vygotsky,et al. Tool and symbol in child development , 2008 .
[86] Steven M. LaValle,et al. Planning algorithms , 2006 .
[87] Sridhar Mahadevan,et al. Recent Advances in Hierarchical Reinforcement Learning , 2003, Discret. Event Dyn. Syst..
[88] Sridhar Mahadevan,et al. Recent Advances in Hierarchical Reinforcement Learning , 2003, Discret. Event Dyn. Syst..
[89] Hector Muñoz-Avila,et al. SHOP: Simple Hierarchical Ordered Planner , 1999, IJCAI.
[90] Peter Carruthers,et al. Thinking in Language?: Evolution and a Modularist Possibility , 1998 .
[91] D. Laplane. Thought and language. , 1992, Behavioural neurology.
[92] Earl David Sacerdoti,et al. A Structure for Plans and Behavior , 1977 .
[93] Richard Fikes,et al. STRIPS: A New Approach to the Application of Theorem Proving to Problem Solving , 1971, IJCAI.
[94] L. Vygotsky. Play and Its Role in the Mental Development of the Child , 1967 .