暂无分享,去创建一个
[1] Bernd Bischl,et al. OpenML Benchmarking Suites , 2017, 1708.03731.
[2] Peter Dayan,et al. Q-learning , 1992, Machine Learning.
[3] Bernd Bischl,et al. First Investigations on Noisy Model-Based Multi-objective Optimization , 2017, EMO.
[4] Joshua B. Tenenbaum,et al. Hierarchical Deep Reinforcement Learning: Integrating Temporal Abstraction and Intrinsic Motivation , 2016, NIPS.
[5] Juliana Freire,et al. AlphaD3M: Machine Learning Pipeline Synthesis , 2021, ArXiv.
[6] Kevin Leyton-Brown,et al. Sequential Model-Based Optimization for General Algorithm Configuration , 2011, LION.
[7] Fangkai Yang,et al. Program Search for Machine Learning Pipelines Leveraging Symbolic Planning and Reinforcement Learning , 2018, GPTP.
[8] Nidhi Kushwaha,et al. A Lesson learned from PMF based approach for Semantic Recommender System , 2017, Journal of Intelligent Information Systems.
[9] Bernd Bischl,et al. OpenML Benchmarking Suites and the OpenML100 , 2017, ArXiv.
[10] Eyke Hüllermeier,et al. ML-Plan: Automated machine learning via hierarchical planning , 2018, Machine Learning.
[11] Quoc V. Le,et al. Neural Architecture Search with Reinforcement Learning , 2016, ICLR.
[12] Wolfgang Banzhaf,et al. Genetic Programming: An Introduction , 1997 .
[13] Bernd Bischl,et al. High Dimensional Restrictive Federated Model Selection with multi-objective Bayesian Optimization over shifted distributions , 2019, IntelliSys.
[14] Gaël Varoquaux,et al. Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..
[15] Bernd Bischl,et al. mlrMBO: A Modular Framework for Model-Based Optimization of Expensive Black-Box Functions , 2017, 1703.03373.
[16] Sridhar Mahadevan,et al. Recent Advances in Hierarchical Reinforcement Learning , 2003, Discret. Event Dyn. Syst..
[17] Demis Hassabis,et al. Mastering the game of Go without human knowledge , 2017, Nature.
[18] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[19] Fangkai Yang,et al. PEORL: Integrating Symbolic Planning and Hierarchical Reinforcement Learning for Robust Decision-Making , 2018, IJCAI.
[20] Kevin Leyton-Brown,et al. Auto-WEKA: Automated Selection and Hyper-Parameter Optimization of Classification Algorithms , 2012, ArXiv.
[21] Gisele L. Pappa,et al. RECIPE: A Grammar-Based Framework for Automatically Evolving Classification Pipelines , 2017, EuroGP.
[22] Aaron Klein,et al. Efficient and Robust Automated Machine Learning , 2015, NIPS.
[23] Bernd Bischl,et al. mlr: Machine Learning in R , 2016, J. Mach. Learn. Res..
[24] Yang Su,et al. Detecting Chinese calligraphy style consistency by deep learning and one-class SVM , 2017, 2017 2nd International Conference on Image, Vision and Computing (ICIVC).
[25] Nando de Freitas,et al. A Tutorial on Bayesian Optimization of Expensive Cost Functions, with Application to Active User Modeling and Hierarchical Reinforcement Learning , 2010, ArXiv.
[26] Ramesh Raskar,et al. Designing Neural Network Architectures using Reinforcement Learning , 2016, ICLR.
[27] Thomas G. Dietterich. Hierarchical Reinforcement Learning with the MAXQ Value Function Decomposition , 1999, J. Artif. Intell. Res..
[28] Yoshua Bengio,et al. Algorithms for Hyper-Parameter Optimization , 2011, NIPS.
[29] Yoshua Bengio,et al. Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..
[30] Randal S. Olson,et al. TPOT: A Tree-based Pipeline Optimization Tool for Automating Machine Learning , 2016, AutoML@ICML.
[31] David D. Cox,et al. Hyperopt: A Python Library for Optimizing the Hyperparameters of Machine Learning Algorithms , 2013, SciPy.
[32] Ameet Talwalkar,et al. Hyperband: A Novel Bandit-Based Approach to Hyperparameter Optimization , 2016, J. Mach. Learn. Res..