暂无分享,去创建一个
[1] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[2] John Schulman,et al. Concrete Problems in AI Safety , 2016, ArXiv.
[3] Francis P. Hunkins,et al. Curriculum: Foundations, Principles, and Issues , 1988 .
[4] P. Dayan,et al. Flexible shaping: How learning in small steps helps , 2009, Cognition.
[5] Ricardo Vilalta,et al. Introduction to the Special Issue on Meta-Learning , 2004, Machine Learning.
[6] Yann LeCun,et al. The Loss Surfaces of Multilayer Networks , 2014, AISTATS.
[7] Alan Fern,et al. A Bayesian Approach for Policy Learning from Trajectory Preference Queries , 2012, NIPS.
[8] Jieyu Zhao,et al. Simple Principles of Metalearning , 1996 .
[9] Surya Ganguli,et al. An analytic theory of generalization dynamics and transfer learning in deep linear networks , 2018, ICLR.
[10] Hugo Larochelle,et al. Optimization as a Model for Few-Shot Learning , 2016, ICLR.
[11] D. G. Stork,et al. Is backpropagation biologically plausible? , 1989, International 1989 Joint Conference on Neural Networks.
[12] Surya Ganguli,et al. Exact solutions to the nonlinear dynamics of learning in deep linear neural networks , 2013, ICLR.
[13] Sebastian Thrun,et al. Lifelong Learning Algorithms , 1998, Learning to Learn.
[14] T. Hensch. Critical period regulation. , 2004, Annual review of neuroscience.
[15] Konrad P. Körding,et al. Toward an Integration of Deep Learning and Neuroscience , 2016, bioRxiv.
[16] Peter L. Bartlett,et al. RL$^2$: Fast Reinforcement Learning via Slow Reinforcement Learning , 2016, ArXiv.
[17] Larry A. Rendell,et al. Layered Concept-Learning and Dynamically Variable Bias Management , 1987, IJCAI.
[18] Roger C. Schank,et al. Conceptual dependency: A theory of natural language understanding , 1972 .
[19] Zhi-Qin John Xu,et al. Training behavior of deep neural network in frequency domain , 2018, ICONIP.
[20] J. M. Davis. Imitation: A Review and Critique , 1973 .
[21] Hao Li,et al. Visualizing the Loss Landscape of Neural Nets , 2017, NeurIPS.
[22] Shane Legg,et al. Deep Reinforcement Learning from Human Preferences , 2017, NIPS.
[23] Oriol Vinyals,et al. Matching Networks for One Shot Learning , 2016, NIPS.
[24] David D. Cox,et al. On the information bottleneck theory of deep learning , 2018, ICLR.
[25] Joshua B. Tenenbaum,et al. Building machines that learn and think like people , 2016, Behavioral and Brain Sciences.
[26] Surya Ganguli,et al. Learning Dynamics of Deep Networks Admit Low-Rank Tensor Descriptions , 2018 .
[27] Terrence J. Sejnowski,et al. Unsupervised Learning , 2018, Encyclopedia of GIS.
[28] R. J. Williams,et al. Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.
[29] Surya Ganguli,et al. On the Expressive Power of Deep Neural Networks , 2016, ICML.
[30] Ron Meir,et al. Meta-Learning by Adjusting Priors Based on Extended PAC-Bayes Theory , 2017, ICML.
[31] Daphna Weinshall,et al. Theory of Curriculum Learning, with Convex Loss Functions , 2018, ArXiv.
[32] Stefan Schaal,et al. Is imitation learning the route to humanoid robots? , 1999, Trends in Cognitive Sciences.
[33] Jason Weston,et al. Curriculum learning , 2009, ICML '09.
[34] Yann LeCun,et al. Comparing dynamics: deep neural networks versus glassy systems , 2018, ICML.
[35] Benjamin Van Roy,et al. Deep Exploration via Bootstrapped DQN , 2016, NIPS.
[36] Stephen P. Boyd,et al. A Differential Equation for Modeling Nesterov's Accelerated Gradient Method: Theory and Insights , 2014, J. Mach. Learn. Res..
[37] Yoshua Bengio,et al. An Empirical Study of Example Forgetting during Deep Neural Network Learning , 2018, ICLR.
[38] Francis Crick,et al. The recent excitement about neural networks , 1989, Nature.
[39] Razvan Pascanu,et al. Ray Interference: a Source of Plateaus in Deep Reinforcement Learning , 2019, ArXiv.
[40] Andrew Y. Ng,et al. Policy Invariance Under Reward Transformations: Theory and Application to Reward Shaping , 1999, ICML.
[41] Filip De Turck,et al. VIME: Variational Information Maximizing Exploration , 2016, NIPS.
[42] Burr Settles,et al. Active Learning , 2012, Synthesis Lectures on Artificial Intelligence and Machine Learning.
[43] Zhi-Qin John Xu,et al. Understanding training and generalization in deep learning by Fourier analysis , 2018, ArXiv.
[44] Lutz Prechelt,et al. Early Stopping - But When? , 2012, Neural Networks: Tricks of the Trade.
[45] Kunihiko Fukushima,et al. Neocognitron: A Self-Organizing Neural Network Model for a Mechanism of Visual Pattern Recognition , 1982 .
[46] Christopher M. Bishop,et al. Pattern Recognition and Machine Learning (Information Science and Statistics) , 2006 .
[47] Bartunov Sergey,et al. Meta-Learning with Memory-Augmented Neural Networks , 2016 .
[48] Surya Ganguli,et al. The Emergence of Spectral Universality in Deep Networks , 2018, AISTATS.
[49] David A. Cohn,et al. Active Learning with Statistical Models , 1996, NIPS.
[50] J. Elman. Learning and development in neural networks: the importance of starting small , 1993, Cognition.
[51] Stuart J. Russell,et al. Research Priorities for Robust and Beneficial Artificial Intelligence , 2015, AI Mag..
[52] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[53] Tom Schaul,et al. Meta-learning by the Baldwin effect , 2018, GECCO.
[54] Marcel A. J. van Gerven,et al. Deep Neural Networks Reveal a Gradient in the Complexity of Neural Representations across the Ventral Stream , 2014, The Journal of Neuroscience.
[55] Colin J. Akerman,et al. Random synaptic feedback weights support error backpropagation for deep learning , 2016, Nature Communications.
[56] Peter Auer,et al. Near-optimal Regret Bounds for Reinforcement Learning , 2008, J. Mach. Learn. Res..
[57] Geoffrey E. Hinton,et al. How Learning Can Guide Evolution , 1996, Complex Syst..
[58] Geoffrey E. Hinton,et al. Learning and relearning in Boltzmann machines , 1986 .
[59] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[60] Stefano Soatto,et al. Stochastic Gradient Descent Performs Variational Inference, Converges to Limit Cycles for Deep Networks , 2017, 2018 Information Theory and Applications Workshop (ITA).
[61] Sebastian Thrun,et al. Learning to Learn: Introduction and Overview , 1998, Learning to Learn.
[62] Joshua B. Tenenbaum,et al. Human-level concept learning through probabilistic program induction , 2015, Science.
[63] James L. McClelland,et al. Semantic Cognition: A Parallel Distributed Processing Approach , 2004 .
[64] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[65] Andre Wibisono,et al. A variational perspective on accelerated methods in optimization , 2016, Proceedings of the National Academy of Sciences.
[66] J. Piaget. Play, dreams and imitation in childhood , 1951 .
[67] Jason J. Corso,et al. A Critical Investigation of Deep Reinforcement Learning for Navigation , 2018, ArXiv.
[68] Surya Ganguli,et al. Task-Driven Convolutional Recurrent Models of the Visual System , 2018, NeurIPS.
[69] Joseph L. Austerweil,et al. People Teach With Rewards and Punishments as Communication, Not Reinforcements , 2019, Journal of experimental psychology. General.
[70] Terence D. Sanger,et al. Neural network learning control of robot manipulators using gradually increasing task difficulty , 1994, IEEE Trans. Robotics Autom..
[71] Shane Legg,et al. Noisy Networks for Exploration , 2017, ICLR.
[72] Jitendra Malik,et al. Learning to Optimize , 2016, ICLR.
[73] K. Plunkett,et al. Connectionism and developmental theory , 1992 .
[74] Daphne Koller,et al. Self-Paced Learning for Latent Variable Models , 2010, NIPS.
[75] G. Peterson. A day of great illumination: B. F. Skinner's discovery of shaping. , 2004, Journal of the experimental analysis of behavior.
[76] Thomas L. Griffiths,et al. Recasting Gradient-Based Meta-Learning as Hierarchical Bayes , 2018, ICLR.
[77] Nathan Srebro,et al. The Implicit Bias of Gradient Descent on Separable Data , 2017, J. Mach. Learn. Res..
[78] Koray Kavukcuoglu,et al. Neural scene representation and rendering , 2018, Science.
[79] Mark O. Riedl,et al. Explore, Exploit or Listen: Combining Human Feedback and Policy Model to Speed up Deep Reinforcement Learning in 3D Worlds , 2017, ArXiv.
[80] William A. Gale,et al. A sequential algorithm for training text classifiers , 1994, SIGIR '94.
[81] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[82] Nikolaus Kriegeskorte,et al. Deep Supervised, but Not Unsupervised, Models May Explain IT Cortical Representation , 2014, PLoS Comput. Biol..
[83] David J. Field,et al. Emergence of simple-cell receptive field properties by learning a sparse code for natural images , 1996, Nature.
[84] Razvan Pascanu,et al. Meta-Learning with Latent Embedding Optimization , 2018, ICLR.
[85] Naftali Tishby,et al. Opening the Black Box of Deep Neural Networks via Information , 2017, ArXiv.
[86] Andrew M. Saxe,et al. High-dimensional dynamics of generalization error in neural networks , 2017, Neural Networks.
[87] Ingmar Kanitscheider,et al. Training recurrent networks to generate hypotheses about how the brain solves hard navigation problems , 2016, NIPS.
[88] Yoshua Bengio,et al. A Closer Look at Memorization in Deep Networks , 2017, ICML.
[89] Pieter Abbeel,et al. Apprenticeship learning via inverse reinforcement learning , 2004, ICML.
[90] Abhinav Gupta,et al. Training Region-Based Object Detectors with Online Hard Example Mining , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[91] J. Galef. • IMITATION IN ANIMALS: HISTORY, DEFINITION, AND INTERPRETATION OF DATA FROM THE PSYCHOLOGICAL LABORATORY , 2013 .
[92] Zeb Kurth-Nelson,et al. Learning to reinforcement learn , 2016, CogSci.
[93] Jonathan Baxter,et al. Theoretical Models of Learning to Learn , 1998, Learning to Learn.
[94] Yong Jae Lee,et al. Learning the easy things first: Self-paced visual category discovery , 2011, CVPR 2011.
[95] Stefan Schaal,et al. Computational approaches to motor learning by imitation. , 2003, Philosophical transactions of the Royal Society of London. Series B, Biological sciences.
[96] Surya Ganguli,et al. A mathematical theory of semantic development in deep neural networks , 2018, Proceedings of the National Academy of Sciences.
[97] Léon Bottou,et al. Large-Scale Machine Learning with Stochastic Gradient Descent , 2010, COMPSTAT.
[98] Jonathan Baxter,et al. A Model of Inductive Bias Learning , 2000, J. Artif. Intell. Res..
[99] Yoshua Bengio,et al. On the Spectral Bias of Deep Neural Networks , 2018, ArXiv.
[100] Vladimir Braverman,et al. The Physical Systems Behind Optimization Algorithms , 2018, NeurIPS.
[101] Oriol Vinyals,et al. Qualitatively characterizing neural network optimization problems , 2014, ICLR.
[102] Pieter Abbeel,et al. Meta-Learning with Temporal Convolutions , 2017, ArXiv.
[103] Stefano Soatto,et al. Emergence of Invariance and Disentanglement in Deep Representations , 2017, 2018 Information Theory and Applications Workshop (ITA).
[104] F ROSENBLATT,et al. The perceptron: a probabilistic model for information storage and organization in the brain. , 1958, Psychological review.
[105] Surya Ganguli,et al. Deep Information Propagation , 2016, ICLR.
[106] Shane Legg,et al. Scalable agent alignment via reward modeling: a research direction , 2018, ArXiv.
[107] Stefano Soatto,et al. Critical Learning Periods in Deep Networks , 2018, ICLR.