暂无分享,去创建一个
[1] Yoshua Bengio,et al. Algorithms for Hyper-Parameter Optimization , 2011, NIPS.
[2] Yoshua Bengio,et al. Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..
[3] Carlos Soares,et al. Ranking Learning Algorithms: Using IBL and Meta-Learning on Accuracy and Time Results , 2003, Machine Learning.
[4] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[5] Marcin Andrychowicz,et al. Learning to learn by gradient descent by gradient descent , 2016, NIPS.
[6] Jasper Snoek,et al. Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.
[7] Nicol N. Schraudolph,et al. 3D hand tracking by rapid stochastic gradient descent using a skinning model , 2004 .
[8] Ricardo Vilalta,et al. Metalearning - Applications to Data Mining , 2008, Cognitive Technologies.
[9] Jürgen Schmidhuber,et al. Optimal Ordered Problem Solver , 2002, Machine Learning.
[10] Ian R. Fasel,et al. Optimization on a Budget: A Reinforcement Learning Approach , 2008, NIPS.
[11] Yann LeCun,et al. Learning Fast Approximations of Sparse Coding , 2010, ICML.
[12] Nichael Lynn Cramer,et al. A Representation for the Adaptive Generation of Simple Sequential Programs , 1985, ICGA.
[13] Sebastian Nowozin,et al. Learning Step Size Controllers for Robust Neural Network Training , 2016, AAAI.
[14] Sergey Levine,et al. End-to-End Training of Deep Visuomotor Policies , 2015, J. Mach. Learn. Res..
[15] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[16] C A Nelson,et al. Learning to Learn , 2017, Encyclopedia of Machine Learning and Data Mining.
[17] Yoram Singer,et al. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..
[18] Frank Hutter,et al. Initializing Bayesian Hyperparameter Optimization via Meta-Learning , 2015, AAAI.
[19] Tat-Seng Chua,et al. Deep Q-Networks for Accelerating the Training of Deep Neural Networks , 2016, ArXiv.
[20] Yoshua Bengio,et al. Gradient-Based Optimization of Hyperparameters , 2000, Neural Computation.
[21] Justin Domke,et al. Generic Methods for Optimization-Based Modeling , 2012, AISTATS.
[22] Samantha Hansen,et al. Using Deep Q-Learning to Control Optimization Hyperparameters , 2016, ArXiv.
[23] Jitendra Malik,et al. Learning to Optimize , 2016, ICLR.
[24] Alex Graves,et al. Neural Turing Machines , 2014, ArXiv.
[25] Kevin Leyton-Brown,et al. Sequential Model-Based Optimization for General Algorithm Configuration , 2011, LION.
[26] Michael I. Jordan,et al. Learning Programs: A Hierarchical Bayesian Approach , 2010, ICML.
[27] Ricardo Vilalta,et al. A Perspective View and Survey of Meta-Learning , 2002, Artificial Intelligence Review.
[28] Sepp Hochreiter,et al. Learning to Learn Using Gradient Descent , 2001, ICANN.
[29] Arindam Banerjee,et al. Bregman Alternating Direction Method of Multipliers , 2013, NIPS.
[30] Jasper Snoek,et al. Multi-Task Bayesian Optimization , 2013, NIPS.
[31] Yoshua Bengio,et al. Learning a synaptic learning rule , 1991, IJCNN-91-Seattle International Joint Conference on Neural Networks.
[32] Guillermo Sapiro,et al. Supervised Sparse Analysis and Synthesis Operators , 2013, NIPS.
[33] Ryan P. Adams,et al. Gradient-based Hyperparameter Optimization through Reversible Learning , 2015, ICML.