A NEW FRAMEWORK FOR PARALLEL RANKING & SELECTION USING AN ADAPTIVE STANDARD

When we have sufficient computational resources to treat a simulation optimization problem as a ranking & selection (R&S) problem, then it can be "solved." R&S is exhaustive search—all feasible solutions are simulated—with meaningful statistical error control. High-performance parallel computing promises to extend the R&S limit to even larger problems, but parallelizing R&S procedures in a way that maintains statistical validity while achieving substantial speed-up is difficult. In this paper we introduce an entirely new framework for R&S called Parallel Adaptive Survivor Selection (PASS) that is specifically engineered to exploit parallel computing environments for solving simulation optimization problems with a very large number of feasible solutions.

[1]  Isaac Dialsingh,et al.  Large-scale inference: empirical Bayes methods for estimation, testing, and prediction , 2012 .

[2]  Barry L. Nelson,et al.  Improving the Efficiency and Efficacy of Controlled Sequential Bifurcation for Simulation Factor Screening , 2010, INFORMS J. Comput..

[3]  Shie Mannor,et al.  PAC Bounds for Multi-armed Bandit and Markov Decision Processes , 2002, COLT.

[4]  Susan R. Hunter,et al.  Parallel Ranking and Selection , 2017 .

[5]  Dashi I. Singham,et al.  Multiple comparisons with a standard using false discovery rates , 2016, 2016 Winter Simulation Conference (WSC).

[6]  Bradley Efron,et al.  Large-scale inference , 2010 .

[7]  Dragan Cvetković,et al.  Modeling and Computer Simulation , 2019 .

[8]  Barry L. Nelson,et al.  Selecting the best system: theory and methods , 2003, Proceedings of the 2003 Winter Simulation Conference, 2003..

[9]  Barry L. Nelson,et al.  Indifference-Zone-Free Selection of the Best , 2016, Oper. Res..

[10]  Susan R. Hunter,et al.  Ranking and selection in a high performance computing environment , 2013, 2013 Winter Simulations Conference (WSC).

[11]  Enver Yücesan,et al.  A new perspective on feasibility determination , 2008, 2008 Winter Simulation Conference.

[12]  Peter I. Frazier,et al.  Sequential Bayes-Optimal Policies for Multiple Comparisons with a Known Standard , 2013, Oper. Res..

[13]  Y. Benjamini,et al.  Controlling the false discovery rate: a practical and powerful approach to multiple testing , 1995 .

[14]  Barry L. Nelson,et al.  Comparisons with a Standard in Simulation Experiments , 2001, Manag. Sci..