暂无分享,去创建一个
[1] David D. Cox,et al. Making a Science of Model Search: Hyperparameter Optimization in Hundreds of Dimensions for Vision Architectures , 2013, ICML.
[2] Hao Wang,et al. A new acquisition function for Bayesian optimization based on the moment-generating function , 2017, 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC).
[3] Yoshua Bengio,et al. Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..
[4] Antanas Zilinskas,et al. A review of statistical models for global optimization , 1992, J. Glob. Optim..
[5] Hao Wang,et al. Multi-point Efficient Global Optimization Using Niching Evolution Strategy , 2015, EVOLVE.
[6] Peter J. Angeline,et al. An evolutionary algorithm that constructs recurrent neural networks , 1994, IEEE Trans. Neural Networks.
[7] Bill C White,et al. Optimization of neural network architecture using genetic programming improves detection and modeling of gene-gene interactions in studies of human diseases , 2003, BMC Bioinformatics.
[8] Peter M. Todd,et al. Designing Neural Networks using Genetic Algorithms , 1989, ICGA.
[9] Thomas Bäck,et al. Mixed Integer Evolution Strategies for Parameter Optimization , 2013, Evolutionary Computation.
[10] M. D. McKay,et al. A comparison of three methods for selecting values of input variables in the analysis of output from a computer code , 2000 .
[11] Le Song,et al. Deep Fried Convnets , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).
[12] Kevin Leyton-Brown,et al. Sequential Model-Based Optimization for General Algorithm Configuration , 2011, LION.
[13] Donald R. Jones,et al. Efficient Global Optimization of Expensive Black-Box Functions , 1998, J. Glob. Optim..
[14] Jasper Snoek,et al. Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.
[15] Frank Hutter,et al. CMA-ES for Hyperparameter Optimization of Deep Neural Networks , 2016, ArXiv.
[16] D. Krige. A statistical approach to some basic mine valuation problems on the Witwatersrand, by D.G. Krige, published in the Journal, December 1951 : introduction by the author , 1951 .
[17] Richard J. Beckman,et al. A Comparison of Three Methods for Selecting Values of Input Variables in the Analysis of Output From a Computer Code , 2000, Technometrics.
[18] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[19] Peter Auer,et al. Using Confidence Bounds for Exploitation-Exploration Trade-offs , 2003, J. Mach. Learn. Res..
[20] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[21] Donald R. Jones,et al. A Taxonomy of Global Optimization Methods Based on Response Surfaces , 2001, J. Glob. Optim..
[22] Andreas Krause,et al. Information-Theoretic Regret Bounds for Gaussian Process Optimization in the Bandit Setting , 2009, IEEE Transactions on Information Theory.
[23] J. Mockus. Bayesian Approach to Global Optimization: Theory and Applications , 1989 .
[24] Kevin Leyton-Brown,et al. Parallel Algorithm Configuration , 2012, LION.
[25] Risto Miikkulainen,et al. Evolving Neural Networks through Augmenting Topologies , 2002, Evolutionary Computation.
[26] Rob Fergus,et al. Stochastic Pooling for Regularization of Deep Convolutional Neural Networks , 2013, ICLR.
[27] Benjamin Graham,et al. Fractional Max-Pooling , 2014, ArXiv.
[28] Jonas Mockus,et al. On Bayesian Methods for Seeking the Extremum , 1974, Optimization Techniques.
[29] Jürgen Schmidhuber,et al. Multi-column deep neural networks for image classification , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.
[30] D. Ginsbourger,et al. Kriging is well-suited to parallelize optimization , 2010 .
[31] Sonja Kuhnt,et al. Design and analysis of computer experiments , 2010 .
[32] Thomas Brox,et al. Striving for Simplicity: The All Convolutional Net , 2014, ICLR.