暂无分享,去创建一个
[1] Kurt Hornik,et al. Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.
[2] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[3] D. Fudenberg,et al. The Theory of Learning in Games , 1998 .
[4] D. Lizotte. Practical bayesian optimization , 2008 .
[5] Carl E. Rasmussen,et al. Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.
[6] Geoffrey E. Hinton,et al. Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.
[7] Tobias Scheffer,et al. Stackelberg games for adversarial prediction problems , 2011, KDD.
[8] Yoshua Bengio,et al. Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..
[9] Jasper Snoek,et al. Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.
[10] Justin Domke,et al. Generic Methods for Optimization-Based Modeling , 2012, AISTATS.
[11] Jasper Snoek,et al. Freeze-Thaw Bayesian Optimization , 2014, ArXiv.
[12] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[13] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[14] Julien Cornebise,et al. Weight Uncertainty in Neural Networks , 2015, ArXiv.
[15] Ryan P. Adams,et al. Gradient-based Hyperparameter Optimization through Reversible Learning , 2015, ICML.
[16] Fabian Pedregosa,et al. Hyperparameter optimization with approximate gradient , 2016, ICML.
[17] Ameet Talwalkar,et al. Efficient Hyperparameter Optimization and Infinitely Many Armed Bandits , 2016, ArXiv.
[18] Tat-Seng Chua,et al. Distilling Reverse-Mode Automatic Differentiation (DrMAD) for Optimizing Hyperparameters of Deep Neural Networks , 2016 .
[19] Kian Hsiang Low,et al. DrMAD: Distilling Reverse-Mode Automatic Differentiation for Optimizing Hyperparameters of Deep Neural Networks , 2016, IJCAI.
[20] Tapani Raiko,et al. Scalable Gradient-Based Tuning of Continuous Regularization Hyperparameters , 2015, ICML.
[21] Quoc V. Le,et al. HyperNetworks , 2016, ICLR.
[22] Ameet Talwalkar,et al. Hyperband: A Novel Bandit-Based Approach to Hyperparameter Optimization , 2016, J. Mach. Learn. Res..
[23] N. Simon,et al. Gradient-based Regularization Parameter Selection for Problems With Nonsmooth Penalty Functions , 2017, 1703.09813.
[24] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[25] Paolo Frasconi,et al. Forward and Reverse Gradient-Based Hyperparameter Optimization , 2017, ICML.
[26] Theodore Lim,et al. SMASH: One-Shot Model Architecture Search through HyperNetworks , 2017, ICLR.