Deep Reinforcement Learning with Model-Based Acceleration for Hyperparameter Optimization

Hyperparameter optimization is a key part of AutoML. In recent years, there have been successful hyperparameter optimization algorithms. However, these methods still face several challenges, such as high cost of evaluating large models or large datasets. In this paper, we introduce a new deep reinforcement learning architecture with model-based acceleration to optimize hyperparameters for any machine learning model. In this method, an agent constructed by a Long Short-Term Memory Network aims at maximizing the expected accuracy of a machine learning model on a validation set. To speed up training, we employ a model to predict the accuracy on a validation set instead of evaluating a machine learning model. To effectively train the agent and the predictive model, Real-Predictive-Real training process is proposed. Besides, to reduce the variance, we propose a bootstrap pool to guide the exploration in the search space. The experiment was carried out by optimizing hyperparameters of two widely used machine learning models: Random Forests and XGBoost. Experimental results show that the proposed method outperforms random search, Bayesian optimization, and Tree-structured Parzen Estimator in terms of accuracy, time efficiency and stability.

[1]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[2]  Marcin Andrychowicz,et al.  Parameter Space Noise for Exploration , 2017, ICLR.

[3]  Ron Kohavi,et al.  Automatic Parameter Selection by Minimizing Estimated Error , 1995, ICML.

[4]  Quoc V. Le,et al.  Neural Architecture Search with Reinforcement Learning , 2016, ICLR.

[5]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[6]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[7]  Senén Barro,et al.  Do we need hundreds of classifiers to solve real world classification problems? , 2014, J. Mach. Learn. Res..

[8]  Prabhat,et al.  Scalable Bayesian Optimization Using Deep Neural Networks , 2015, ICML.

[9]  Ronald J. Williams,et al.  Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.

[10]  Kevin Leyton-Brown,et al.  Sequential Model-Based Optimization for General Algorithm Configuration , 2011, LION.

[11]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[12]  David J. Spiegelhalter,et al.  Machine Learning, Neural and Statistical Classification , 2009 .

[13]  Jasper Snoek,et al.  Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.

[14]  Li Li,et al.  eXtreme Gradient Boosting for Identifying Individual Users Across Different Digital Devices , 2016, WAIM.

[15]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[16]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[17]  Yoshua Bengio,et al.  Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..