暂无分享,去创建一个
[1] Kevin Leyton-Brown,et al. Auto-WEKA: combined selection and hyperparameter optimization of classification algorithms , 2012, KDD.
[2] Alexander Allen,et al. Benchmarking Automatic Machine Learning Frameworks , 2018, ArXiv.
[3] Hilan Bensusan,et al. Meta-Learning by Landmarking Various Learning Algorithms , 2000, ICML.
[4] Francesca Mangili,et al. Should We Really Use Post-Hoc Tests Based on Mean-Ranks? , 2015, J. Mach. Learn. Res..
[5] Kevin Leyton-Brown,et al. Sequential Model-Based Optimization for General Algorithm Configuration , 2011, LION.
[6] R. E. Lee,et al. Distribution-free multiple comparisons between successive treatments , 1995 .
[7] Aaron Klein,et al. Efficient and Robust Automated Machine Learning , 2015, NIPS.
[8] Frank Hutter,et al. Initializing Bayesian Hyperparameter Optimization via Meta-Learning , 2015, AAAI.
[9] Luís Torgo,et al. OpenML: networked science in machine learning , 2014, SKDD.
[10] Fela Winkelmolen,et al. A simple transfer-learning extension of Hyperband , 2018 .
[11] Ameet Talwalkar,et al. Massively Parallel Hyperparameter Tuning , 2018, ArXiv.
[12] Borja Calvo,et al. scmamp: Statistical Comparison of Multiple Algorithms in Multiple Problems , 2016, R J..
[13] Yoshua Bengio,et al. Algorithms for Hyper-Parameter Optimization , 2011, NIPS.
[14] Yoshua Bengio,et al. Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..
[15] Janez Demsar,et al. Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..
[16] Robert D. Nowak,et al. Best-arm identification algorithms for multi-armed bandits in the fixed confidence setting , 2014, 2014 48th Annual Conference on Information Sciences and Systems (CISS).
[17] Madeleine Udell,et al. OBOE: Collaborative Filtering for AutoML Initialization. , 2018 .
[18] Jost Tobias Springenberg,et al. Supplementary material for : Initializing Bayesian Hyperparameter Optimization via Meta-Learning , 2015 .
[19] Randal S. Olson,et al. TPOT: A Tree-based Pipeline Optimization Tool for Automating Machine Learning , 2016, AutoML@ICML.
[20] Peter A. Flach,et al. Advances in Neural Information Processing Systems 28 , 2015 .
[21] Ameet Talwalkar,et al. Efficient Hyperparameter Optimization and Infinitely Many Armed Bandits , 2016, ArXiv.
[22] Aaron Klein,et al. BOHB: Robust and Efficient Hyperparameter Optimization at Scale , 2018, ICML.
[23] S. García,et al. An Extension on "Statistical Comparisons of Classifiers over Multiple Data Sets" for all Pairwise Comparisons , 2008 .
[24] Ameet Talwalkar,et al. Non-stochastic Best Arm Identification and Hyperparameter Optimization , 2015, AISTATS.
[25] Francisco Herrera,et al. Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: Experimental analysis of power , 2010, Inf. Sci..
[26] Thomas G. Dietterich. Approximate Statistical Tests for Comparing Supervised Classification Learning Algorithms , 1998, Neural Computation.
[27] Ameet Talwalkar,et al. Random Search and Reproducibility for Neural Architecture Search , 2019, UAI.
[28] Dimitrios Sarigiannis,et al. Learning to Tune XGBoost with XGBoost , 2019, ArXiv.