暂无分享,去创建一个
[1] Tom Schaul,et al. No more pesky learning rates , 2012, ICML.
[2] Aurélien Garivier,et al. On Upper-Confidence Bound Policies for Non-Stationary Bandit Problems , 2008, 0805.3415.
[3] Matthew D. Zeiler. ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.
[4] Quoc V. Le,et al. On optimization methods for deep learning , 2011, ICML.
[5] Yoshua Bengio,et al. Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..
[6] Mark W. Schmidt,et al. A Stochastic Gradient Method with an Exponential Convergence Rate for Finite Training Sets , 2012, NIPS.
[7] Jasper Snoek,et al. Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.
[8] Jürgen Schmidhuber,et al. Learning dynamic algorithm portfolios , 2006, Annals of Mathematics and Artificial Intelligence.