Tell me why! Explanations support learning relational and causal structure
暂无分享,去创建一个
James L. McClelland | Andrew Kyle Lampinen | Nicholas A. Roy | Neil C. Rabinowitz | Stephanie C. Y. Chan | Allison C. Tam | Jane X. Wang | Adam Santoro | Felix Hill | Ishita Dasgupta | Chen Yan | I. Dasgupta
[1] Andrew Kyle Lampinen,et al. Semantic Exploration from Language Abstractions and Pretrained Representations , 2022, ArXiv.
[2] Improving Intrinsic Exploration with Language Abstractions , 2022, ArXiv.
[3] Mohit Bansal,et al. When Can Models Learn From Explanations? A Formal Framework for Understanding the Roles of Explanation Data , 2021, LNLS.
[4] Marcel van Gerven,et al. Explainable Deep Learning: A Field Guide for the Uninitiated , 2020, J. Artif. Intell. Res..
[5] P. Abbeel,et al. Teachable Reinforcement Learning via Advice Distillation , 2022, NeurIPS.
[6] A Survey of Generalisation in Deep Reinforcement Learning , 2021, ArXiv.
[7] Oriol Vinyals,et al. Highly accurate protein structure prediction with AlphaFold , 2021, Nature.
[8] Sergey Levine,et al. Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability , 2021, NeurIPS.
[9] James L. McClelland,et al. What underlies rapid learning and systematic generalization in humans , 2021, ArXiv.
[10] Alan Yuille,et al. Visual analogy: Deep learning versus compositional models , 2021, ArXiv.
[11] Francesca Toni,et al. Explanation-Based Human Debugging of NLP Models: A Survey , 2021, Transactions of the Association for Computational Linguistics.
[12] J. Bowers,et al. Can Deep Convolutional Neural Networks Learn Same-Different Relations? , 2021, bioRxiv.
[13] Andrew Kyle Lampinen,et al. Symbolic Behaviour in Artificial Intelligence , 2021, ArXiv.
[14] K. Holyoak,et al. Emergence of relational reasoning , 2021, Current Opinion in Behavioral Sciences.
[15] A. Wright,et al. Issues in the comparative cognition of same/different abstract-concept learning , 2021, Current Opinion in Behavioral Sciences.
[16] S. Gershman,et al. Memory as a Computational Resource , 2021, Trends in Cognitive Sciences.
[17] K. Kersting,et al. Right for the Right Concept: Revising Neuro-Symbolic Concepts by Interacting with their Explanations , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[18] Mudit Verma,et al. Widening the Pipeline in Human-Guided Reinforcement Learning with Explanation and Context-Aware Data Augmentation , 2020, NeurIPS.
[19] Zeb Kurth-Nelson,et al. Alchemy: A structured task distribution for meta-reinforcement learning , 2021, ArXiv.
[20] Luis C. Lamb,et al. Neurosymbolic AI: the 3rd wave , 2020, Artificial Intelligence Review.
[21] N. Thompson. Metaphysical Explanation , 2020, The Routledge Handbook of Metametaphysics.
[22] Ilya Kostrikov,et al. Automatic Data Augmentation for Generalization in Deep Reinforcement Learning , 2020, ArXiv.
[23] Andrew Kyle Lampinen,et al. What shapes feature representations? Exploring datasets, architectures, and training , 2020, NeurIPS.
[24] Christopher Potts,et al. Relational reasoning and generalization using non-symbolic neural networks , 2020, CogSci.
[25] M. Bethge,et al. Shortcut learning in deep neural networks , 2020, Nature Machine Intelligence.
[26] Gary Marcus,et al. The Next Decade in AI: Four Steps Towards Robust Artificial Intelligence , 2020, ArXiv.
[27] Fabio Viola,et al. Causally Correct Partial Models for Reinforcement Learning , 2020, ArXiv.
[28] Kristian Kersting,et al. Making deep neural networks right for the right scientific reasons by interacting with their explanations , 2020, Nat. Mach. Intell..
[29] Noah D. Goodman,et al. Shaping Visual Representations with Language for Few-Shot Classification , 2019, ACL.
[30] Razvan Pascanu,et al. Stabilizing Transformers for Reinforcement Learning , 2019, ICML.
[31] Oleg O. Sushkov,et al. Scaling data-driven robotics with reward sketching and batch reinforcement learning , 2019, Robotics: Science and Systems.
[32] James L. McClelland,et al. Environmental drivers of systematicity and generalization in a situated agent , 2019, ICLR.
[33] Murray Shanahan,et al. An Explicitly Relational Neural Network Architecture , 2019, ICML.
[34] Francisco S. Melo,et al. Learning from Explanations and Demonstrations: A Pilot Study , 2020, NL4XAI.
[35] Guy Dove. More than a scaffold: Language is a neuroenhancement , 2020, Cognitive neuropsychology.
[36] Aaron van den Oord,et al. Shaping Belief States with Generative Environment Models for RL , 2019, NeurIPS.
[37] Chelsea Finn,et al. Language as an Abstraction for Hierarchical Deep Reinforcement Learning , 2019, NeurIPS.
[38] Shimon Whiteson,et al. A Survey of Reinforcement Learning Informed by Natural Language , 2019, IJCAI.
[39] Manuela Veloso,et al. Generation of Policy-Level Explanations for Reinforcement Learning , 2019, AAAI.
[40] Yee Whye Teh,et al. Meta reinforcement learning as task inference , 2019, ArXiv.
[41] D. Gentner,et al. Explanation recruits comparison in a category-learning task , 2019, Cognition.
[42] Sergey Levine,et al. Efficient Off-Policy Meta-Reinforcement Learning via Probabilistic Context Variables , 2019, ICML.
[43] Prasoon Goyal,et al. Using Natural Language for Reward Shaping in Reinforcement Learning , 2019, IJCAI.
[44] Judea Pearl,et al. The seven tools of causal inference, with reflections on machine learning , 2019, Commun. ACM.
[45] Zeb Kurth-Nelson,et al. Causal Reasoning from Meta-reinforcement Learning , 2019, ArXiv.
[46] Thomas Lukasiewicz,et al. e-SNLI: Natural Language Inference with Natural Language Explanations , 2018, NeurIPS.
[47] Felix Hill,et al. Measuring abstract reasoning in neural networks , 2018, ICML.
[48] Le Song,et al. Learning to Explain: An Information-Theoretic Perspective on Model Interpretation , 2018, ICML.
[49] Shane Legg,et al. IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures , 2018, ICML.
[50] Judea Pearl,et al. Theoretical Impediments to Machine Learning With Seven Sparks from the Causal Revolution , 2018, WSDM.
[51] Dan Klein,et al. Learning with Latent Language , 2017, NAACL.
[52] Demis Hassabis,et al. Grounded Language Learning in a Simulated 3D World , 2017, ArXiv.
[53] Razvan Pascanu,et al. A simple neural network module for relational reasoning , 2017, NIPS.
[54] Chris Sauer,et al. Beating Atari with Natural Language Guided Reinforcement Learning , 2017, ArXiv.
[55] Andrew Slavin Ross,et al. Right for the Right Reasons: Training Differentiable Models by Constraining their Explanations , 2017, IJCAI.
[56] Tom Schaul,et al. Reinforcement Learning with Unsupervised Auxiliary Tasks , 2016, ICLR.
[57] Michael R. Waldmann,et al. The Oxford handbook of causal reasoning , 2017 .
[58] G. Lupyan. The Centrality of Language in Human Cognition , 2016 .
[59] Joshua B. Tenenbaum,et al. Building machines that learn and think like people , 2016, Behavioral and Brain Sciences.
[60] G. Lupyan,et al. What makes words special? Words as unmotivated cues , 2015, Cognition.
[61] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[62] Robert L. Goldstone,et al. Concreteness Fading in Mathematics and Science Instruction: a Systematic Review , 2014 .
[63] Jivko Sinapov,et al. The odd one out task: Toward an intelligence test for robots , 2010, 2010 IEEE 9th International Conference on Development and Learning.
[64] Joseph Jay Williams,et al. The role of explanation in discovery and generalization: evidence from category learning , 2010, ICLS.
[65] E. Warrington,et al. The different representational frameworks underpinning abstract and concrete knowledge: Evidence from odd-one-out judgements , 2009, Quarterly journal of experimental psychology.
[66] Dedre Gentner,et al. Relational language supports relational cognition in humans and apes , 2008, Behavioral and Brain Sciences.
[67] Derek C. Penn,et al. Darwin's mistake: Explaining the discontinuity between human and nonhuman minds , 2008, Behavioral and Brain Sciences.
[68] G. Lupyan. Taking symbols for granted? Is the discontinuity between human and nonhuman minds the product of external symbol systems? , 2008, Behavioral and Brain Sciences.
[69] Daniel J. Navarro,et al. One of these greebles is not like the others: Semi-supervised models for similarity structures , 2008 .
[70] Keith A. Markus,et al. Making Things Happen: A Theory of Causal Explanation , 2007 .
[71] T. Lombrozo. The structure and function of explanations , 2006, Trends in Cognitive Sciences.
[72] S. Carey,et al. Functional explanation and the function of explanation , 2006, Cognition.
[73] B. Rittle-Johnson,et al. Promoting transfer: effects of self-explanation and direct instruction. , 2006, Child development.
[74] Dedre Gentner,et al. Why we’re so smart , 2003 .
[75] D. Gentner,et al. Language in Mind: Advances in the Study of Language and Thought , 2003 .
[76] David M. Sobel,et al. Detecting blickets: how young children use information about novel causal powers in categorization and induction. , 2000, Child development.
[77] Robert A. Wilson,et al. Explanation and Cognition , 2000 .
[78] A. Gopnik,et al. The scientist in the crib : minds, brains, and how children learn , 1999 .
[79] Michelene T. H. Chi,et al. Eliciting Self-Explanations Improves Understanding , 1994, Cogn. Sci..
[80] R. Mooney,et al. Schema acquisition from a single example , 1992 .
[81] J. Fodor,et al. Connectionism and cognitive architecture: A critical analysis , 1988, Cognition.
[82] Raimo Tuomela,et al. A Pragmatic Theory of Explanation , 1984 .
[83] J. Bruner,et al. The role of tutoring in problem solving. , 1976, Journal of child psychology and psychiatry, and allied disciplines.