Evaluating random forest models for irace

Automatic algorithm configurators can greatly improve the performance of algorithms by effectively searching the parameter space. As algorithm configuration tasks can have large parameter spaces and the execution of candidate algorithm configurations is often very costly in terms of computation time, further improvements in the search techniques used by automatic configurators are important and increase the applicability of available configuration methods. One common technique to improve the behavior of search methods when evaluations are computationally expensive are surrogate model techniques. These models are able to exploit the scarce available data and help to direct the search towards evaluating the most promising candidate configurations. In this paper, we study the use of random forests models as surrogate models in irace, a flexible automatic configuration tool based on iterated racing that has been successfully applied in the literature. We evaluate the performance of the random forest model using different settings when trained with data obtained from the irace configuration process and we evaluate their performance under similar conditions as in the configuration process. This preliminary work aims at providing guidelines for the incorporation of random forest to the configuration process of irace.

[1]  Alan J. Hu,et al.  Boosting Verification by Automatic Tuning of Decision Procedures , 2007 .

[2]  Thomas Stützle,et al.  F-Race and Iterated F-Race: An Overview , 2010, Experimental Methods for the Analysis of Optimization Algorithms.

[3]  Thomas Bartz-Beielstein,et al.  The Sequential Parameter Optimization Toolbox , 2010, Experimental Methods for the Analysis of Optimization Algorithms.

[4]  Alan J. Hu,et al.  Boosting Verification by Automatic Tuning of Decision Procedures , 2007, Formal Methods in Computer Aided Design (FMCAD'07).

[5]  Thomas Bartz-Beielstein,et al.  Experimental Methods for the Analysis of Optimization Algorithms , 2010 .

[6]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[7]  Leifur Leifsson,et al.  Surrogate-Based Methods , 2011, Computational Optimization, Methods and Algorithms.

[8]  Leslie Pérez Cáceres,et al.  The irace package: Iterated racing for automatic algorithm configuration , 2016 .

[9]  Yuri Malitsky,et al.  Model-Based Genetic Algorithms for Algorithm Configuration , 2015, IJCAI.

[10]  Thomas Bartz-Beielstein,et al.  Tuning search algorithms for real-world applications: a regression tree based approach , 2004, Proceedings of the 2004 Congress on Evolutionary Computation (IEEE Cat. No.04TH8753).

[11]  Kevin Leyton-Brown,et al.  Algorithm runtime prediction: Methods & evaluation , 2012, Artif. Intell..

[12]  Thomas Stützle,et al.  Improvement Strategies for the F-Race Algorithm: Sampling Design and Iterative Refinement , 2007, Hybrid Metaheuristics.

[13]  Andreas Ziegler,et al.  ranger: A Fast Implementation of Random Forests for High Dimensional Data in C++ and R , 2015, 1508.04409.

[14]  Donald R. Jones,et al.  Efficient Global Optimization of Expensive Black-Box Functions , 1998, J. Glob. Optim..

[15]  Kevin Leyton-Brown,et al.  Sequential Model-Based Optimization for General Algorithm Configuration , 2011, LION.