Automatic Configuration of Deep Neural Networks with Parallel Efficient Global Optimization
暂无分享,去创建一个
[1] Bill C White,et al. Optimization of neural network architecture using genetic programming improves detection and modeling of gene-gene interactions in studies of human diseases , 2003, BMC Bioinformatics.
[2] Kevin Leyton-Brown,et al. Sequential Model-Based Optimization for General Algorithm Configuration , 2011, LION.
[3] Benjamin Graham,et al. Fractional Max-Pooling , 2014, ArXiv.
[4] Jonas Mockus,et al. On Bayesian Methods for Seeking the Extremum , 1974, Optimization Techniques.
[5] Donald R. Jones,et al. Efficient Global Optimization of Expensive Black-Box Functions , 1998, J. Glob. Optim..
[6] Risto Miikkulainen,et al. Evolving Neural Networks through Augmenting Topologies , 2002, Evolutionary Computation.
[7] Rob Fergus,et al. Stochastic Pooling for Regularization of Deep Convolutional Neural Networks , 2013, ICLR.
[8] Jasper Snoek,et al. Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.
[9] J. Mockus. Bayesian Approach to Global Optimization: Theory and Applications , 1989 .
[10] D. Krige. A statistical approach to some basic mine valuation problems on the Witwatersrand, by D.G. Krige, published in the Journal, December 1951 : introduction by the author , 1951 .
[11] Richard J. Beckman,et al. A Comparison of Three Methods for Selecting Values of Input Variables in the Analysis of Output From a Computer Code , 2000, Technometrics.
[12] Hao Wang,et al. A new acquisition function for Bayesian optimization based on the moment-generating function , 2017, 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC).
[13] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[14] Yoshua Bengio,et al. Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..
[15] Peter J. Angeline,et al. An evolutionary algorithm that constructs recurrent neural networks , 1994, IEEE Trans. Neural Networks.
[16] Frank Hutter,et al. CMA-ES for Hyperparameter Optimization of Deep Neural Networks , 2016, ArXiv.
[17] Peter Auer,et al. Using Confidence Bounds for Exploitation-Exploration Trade-offs , 2003, J. Mach. Learn. Res..
[18] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[19] Sonja Kuhnt,et al. Design and analysis of computer experiments , 2010 .
[20] Andreas Krause,et al. Information-Theoretic Regret Bounds for Gaussian Process Optimization in the Bandit Setting , 2009, IEEE Transactions on Information Theory.
[21] Thomas Brox,et al. Striving for Simplicity: The All Convolutional Net , 2014, ICLR.
[22] Hao Wang,et al. Multi-point Efficient Global Optimization Using Niching Evolution Strategy , 2015, EVOLVE.
[23] Antanas Zilinskas,et al. A review of statistical models for global optimization , 1992, J. Glob. Optim..
[24] Peter M. Todd,et al. Designing Neural Networks using Genetic Algorithms , 1989, ICGA.
[25] Le Song,et al. Deep Fried Convnets , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).
[26] Thomas Bäck,et al. Mixed Integer Evolution Strategies for Parameter Optimization , 2013, Evolutionary Computation.
[27] David D. Cox,et al. Making a Science of Model Search: Hyperparameter Optimization in Hundreds of Dimensions for Vision Architectures , 2013, ICML.
[28] Donald R. Jones,et al. A Taxonomy of Global Optimization Methods Based on Response Surfaces , 2001, J. Glob. Optim..
[29] Jürgen Schmidhuber,et al. Multi-column deep neural networks for image classification , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.
[30] D. Ginsbourger,et al. Kriging is well-suited to parallelize optimization , 2010 .
[31] Kevin Leyton-Brown,et al. Parallel Algorithm Configuration , 2012, LION.