Stochastic Optimization: Approximate Bayesian Inference and Complete Expected Improvement

Title of dissertation: STOCHASTIC OPTIMIZATION: APPROXIMATE BAYESIAN INFERENCE AND COMPLETE EXPECTED IMPROVEMENT Ye Chen Doctor of Philosophy, 2018 Dissertation directed by: Professor Ilya Ryzhov Department of Decision, Operations, and Information Technologies Stochastic optimization includes modeling, computing and decision making. In practice, due to the limitation of mathematical tools or real budget, many practical solution methods are designed using approximation techniques or taking forms that are efficient to compute and update. These models have shown their practical benefits in different backgrounds, but many of them also lack rigorous theoretical support. Through interfacing with statistical tools, we analyze the asymptotic properties of two important Bayesian models and show their validity by proving consistency or other limiting results, which may be useful to algorithmic scientists seeking to leverage these computational techniques for their practical performance. The first part of the thesis is the consistency analysis of sequential learning algorithms under approximate Bayesian inference. Approximate Bayesian inference is a powerful methodology for constructing computationally efficient statistical mechanisms for sequential learning from incomplete or censored information.Approximate Bayesian learning models have proven successful in a variety of operations research and business problems; however, prior work in this area has been primarily computational, and the consistency of approximate Bayesian estimators has been a largely open problem. We develop a new consistency theory by interpreting approximate Bayesian inference as a form of stochastic approximation (SA) with an additional “bias” term. We prove the convergence of a general SA algorithm of this form, and leverage this analysis to derive the first consistency proofs for a suite of approximate Bayesian models from the recent literature. The second part of the thesis proposes a budget allocation algorithm for the ranking and selection problem. The ranking and selection problem is a well-known mathematical framework for the formal study of optimal information collection. Expected improvement (EI) is a leading algorithmic approach to this problem; the practical benefits of EI have repeatedly been demonstrated in the literature, especially in the widely studied setting of Gaussian sampling distributions. However, it was recently proved that some of the most well-known EI-type methods achieve suboptimal convergence rates. We investigate a recently-proposed variant of EI (known as “complete EI”) and prove that, with some minor modifications, it can be made to converge to the rate-optimal static budget allocation without requiring any tuning. STOCHASTIC OPTIMIZATION: APPROXIMATE BAYESIAN INFERENCE AND COMPLETE EXPECTED IMPROVEMENT

[1]  H. Robbins A Stochastic Approximation Method , 1951 .

[2]  Daniel Russo,et al.  Simple Bayesian Algorithms for Best Arm Identification , 2016, COLT.

[3]  Stuart J. Russell,et al.  Bayesian Q-Learning , 1998, AAAI/IAAI.

[4]  Barry L. Nelson,et al.  A brief introduction to optimization via simulation , 2009, Proceedings of the 2009 Winter Simulation Conference (WSC).

[5]  Chen Ye,et al.  Approximate Bayesian inference as a form of stochastic approximation: A new consistency theory with applications , 2016 .

[6]  So Young Sohn,et al.  Random effects logistic regression model for default prediction of technology credit guarantee fund , 2007, Eur. J. Oper. Res..

[7]  Qiong Zhang,et al.  Moment-Matching-Based Conjugacy Approximation for Bayesian Ranking and Selection , 2016, ACM Trans. Model. Comput. Simul..

[8]  Michael I. Jordan,et al.  MASSACHUSETTS INSTITUTE OF TECHNOLOGY ARTIFICIAL INTELLIGENCE LABORATORY and CENTER FOR BIOLOGICAL AND COMPUTATIONAL LEARNING DEPARTMENT OF BRAIN AND COGNITIVE SCIENCES , 1996 .

[9]  Chun-Hung Chen,et al.  Simulation Budget Allocation for Further Enhancing the Efficiency of Ordinal Optimization , 2000, Discret. Event Dyn. Syst..

[10]  Chong Wang,et al.  Variational inference in nonconjugate models , 2012, J. Mach. Learn. Res..

[11]  R. Bechhofer A Single-Sample Multiple Decision Procedure for Ranking Means of Normal Populations with known Variances , 1954 .

[12]  John N. Tsitsiklis,et al.  Asynchronous stochastic approximation and Q-learning , 1994, Mach. Learn..

[13]  Thomas Hofmann,et al.  TrueSkill™: A Bayesian Skill Rating System , 2007 .

[14]  Huashuai Qu,et al.  Learning Demand Curves in B2B Pricing: A New Framework and Case Study , 2020 .

[15]  Jean-Michel Marin,et al.  Approximate Bayesian computational methods , 2011, Statistics and Computing.

[16]  S. Jain,et al.  CALIBRATING SIMULATION MODELS USING THE KNOWLEDGE GRADIENT WITH CONTINUOUS PARAMETERS , 2010 .

[17]  Shie Mannor,et al.  Reinforcement learning with Gaussian processes , 2005, ICML.

[18]  Angelia Nedic,et al.  On stochastic gradient and subgradient methods with adaptive steplength sequences , 2011, Autom..

[19]  Barry L. Nelson,et al.  A fully sequential procedure for indifference-zone selection in simulation , 2001, TOMC.

[20]  Loo Hay Lee,et al.  Ranking and Selection: Efficient Simulation Budget Allocation , 2015 .

[21]  Huashuai Qu,et al.  Simulation optimization: A tutorial overview and recent developments in gradient-based methods , 2014, Proceedings of the Winter Simulation Conference 2014.

[22]  Warren B. Powell,et al.  SMART: A Stochastic Multiscale Model for the Analysis of Energy Resources, Technology, and Policy , 2012, INFORMS J. Comput..

[23]  H. Ruben A New Asymptotic Expansion for the Normal Probability Integral and Mill's Ratio , 1962 .

[24]  Michael C. Fu,et al.  Myopic Allocation Policy With Asymptotically Optimal Sampling Rate , 2017, IEEE Transactions on Automatic Control.

[25]  J Jaap Wessels,et al.  Diagnosing order planning performance at a navy maintenance and repair organization, using logistic regression , 2009 .

[26]  Ilya O. Ryzhov,et al.  On the Convergence Rates of Expected Improvement Methods , 2016, Oper. Res..

[27]  Warren B. Powell,et al.  Information Collection on a Graph , 2011, Oper. Res..

[28]  Jürgen Branke,et al.  Selecting a Selection Procedure , 2007, Manag. Sci..

[29]  Ángel F. García-Fernández,et al.  Gaussian MAP Filtering Using Kalman Optimization , 2015, IEEE Transactions on Automatic Control.

[30]  Dimitris Bertsimas,et al.  From Predictive to Prescriptive Analytics , 2014, Manag. Sci..

[31]  Loo Hay Lee,et al.  Stochastically Constrained Ranking and Selection via SCORE , 2014, ACM Trans. Model. Comput. Simul..

[32]  Houyuan Jiang,et al.  Stochastic Approximation Approaches to the Stochastic Variational Inequality Problem , 2008, IEEE Transactions on Automatic Control.

[33]  Qiong Zhang,et al.  Simulation selection for empirical model comparison , 2015, 2015 Winter Simulation Conference (WSC).

[34]  Michael I. Jordan,et al.  Bayesian parameter estimation via variational methods , 2000, Stat. Comput..

[35]  Warren B. Powell,et al.  Approximate Dynamic Programming Captures Fleet Operations for Schneider National , 2010, Interfaces.

[36]  Zheng Wen,et al.  Efficient Exploration and Value Function Generalization in Deterministic Systems , 2013, NIPS.

[37]  Vivek S. Borkar,et al.  Stochastic Approximation for Nonexpansive Maps: Application to Q-Learning Algorithms , 1997, SIAM J. Control. Optim..

[38]  A. Shiryaev,et al.  Probability (2nd ed.) , 1995, Technometrics.

[39]  U. Rieder,et al.  Markov Decision Processes , 2010 .

[40]  H. Kushner,et al.  Stochastic Approximation and Recursive Algorithms and Applications , 2003 .

[41]  L. Bottou Learning and Stochastic Approximations 3 Q ( z , w ) measures the economical cost ( in hard currency units ) of delivering , 2012 .

[42]  Tom Minka,et al.  TrueSkill Through Time: Revisiting the History of Chess , 2007, NIPS.

[43]  Loo Hay Lee,et al.  Stochastic Simulation Optimization - An Optimal Computing Budget Allocation , 2010, System Engineering and Operations Research.

[44]  Sean P. Meyn,et al.  The O.D.E. Method for Convergence of Stochastic Approximation and Reinforcement Learning , 2000, SIAM J. Control. Optim..

[45]  Peter W. Glynn,et al.  A large deviations perspective on ordinal optimization , 2004, Proceedings of the 2004 Winter Simulation Conference, 2004..

[46]  Shie Mannor,et al.  Bayes Meets Bellman: The Gaussian Process Approach to Temporal Difference Learning , 2003, ICML.

[47]  Sanmay Das,et al.  Learning the demand curve in posted-price digital goods auctions , 2011, AAMAS.

[48]  Stephen E. Chick,et al.  Chapter 9 Subjective Probability and Bayesian Methodology , 2006, Simulation.

[49]  Csaba Szepesvári,et al.  The Asymptotic Convergence-Rate of Q-learning , 1997, NIPS.

[50]  Huashuai Qu,et al.  Learning logistic demand curves in business-to-business pricing , 2013, 2013 Winter Simulations Conference (WSC).

[51]  David T. Frazier,et al.  Auxiliary Likelihood-Based Approximate Bayesian Computation in State Space Models , 2016, Journal of Computational and Graphical Statistics.

[52]  Tapabrata Maiti,et al.  Bayesian Data Analysis (2nd ed.) (Book) , 2004 .

[53]  Sanmay Das,et al.  Adapting to a Market Shock: Optimal Sequential Market-Making , 2008, NIPS.

[54]  Angelia Nedic,et al.  Regularized Iterative Stochastic Approximation Methods for Stochastic Variational Inequality Problems , 2013, IEEE Transactions on Automatic Control.

[55]  Warren B. Powell,et al.  Bayesian Exploration for Approximate Dynamic Programming , 2019, Oper. Res..

[56]  Ilya O. Ryzhov,et al.  Approximate Bayesian inference for simulation and optimization , 2015 .

[57]  J. Bather,et al.  Multi‐Armed Bandit Allocation Indices , 1990 .

[58]  Simon Tavaré,et al.  Approximate Bayesian Computation and MCMC , 2004 .

[59]  Xiaoqing Xie,et al.  A Choice‐Based Dynamic Programming Approach for Setting Opaque Prices , 2012 .

[60]  H. Robbins,et al.  A Convergence Theorem for Non Negative Almost Supermartingales and Some Applications , 1985 .

[61]  Ye Chen,et al.  Rate-optimality of the complete expected improvement criterion , 2017, 2017 Winter Simulation Conference (WSC).

[62]  A. Rukhin Matrix Variate Distributions , 1999, The Multivariate Normal Distribution.

[63]  Peter W. Glynn,et al.  A new proof of convergence of MCMC via the ergodic theorem , 2011 .

[64]  Barry L. Nelson,et al.  Discrete optimization via simulation using Gaussian Markov random fields , 2014, Proceedings of the Winter Simulation Conference 2014.

[65]  Christian P. Robert,et al.  On Consistency of Approximate Bayesian Computation , 2015, 1508.05178.

[66]  Susan R. Hunter,et al.  Maximizing quantitative traits in the mating design problem via simulation-based Pareto estimation , 2016 .

[67]  David J. Spiegelhalter,et al.  Sequential updating of conditional probabilities on directed graphical structures , 1990, Networks.

[68]  Sujin Kim,et al.  The stochastic root-finding problem: Overview, solutions, and open questions , 2011, TOMC.

[69]  Warren B. Powell,et al.  Approximate Dynamic Programming - Solving the Curses of Dimensionality , 2007 .

[70]  Sanmay Das,et al.  Instructor Rating Markets , 2013, AAAI.

[71]  Huashuai Qu,et al.  Sequential Selection with Unknown Correlation Structures , 2015, Oper. Res..

[72]  Jukka Corander,et al.  Approximate Bayesian Computation , 2013, PLoS Comput. Biol..

[73]  Tom Minka,et al.  A family of algorithms for approximate Bayesian inference , 2001 .

[74]  M. Degroot Optimal Statistical Decisions , 1970 .

[75]  Boris Defourny,et al.  Optimal Learning in Linear Regression with Combinatorial Feature Selection , 2016, INFORMS J. Comput..

[76]  Vivek F. Farias,et al.  Optimistic Gittins Indices , 2016, NIPS.

[77]  Maqbool Dada,et al.  Pricing and the Newsvendor Problem: A Review with Extensions , 1999, Oper. Res..

[78]  Jürgen Branke,et al.  Sequential Sampling to Myopically Maximize the Expected Value of Information , 2010, INFORMS J. Comput..

[79]  Warren B. Powell,et al.  Optimal Learning: Powell/Optimal , 2012 .

[80]  Donald R. Jones,et al.  Efficient Global Optimization of Expensive Black-Box Functions , 1998, J. Glob. Optim..

[81]  Chong Wang,et al.  Stochastic variational inference , 2012, J. Mach. Learn. Res..

[82]  F. Downton Stochastic Approximation , 1969, Nature.

[83]  H. Haario,et al.  An adaptive Metropolis algorithm , 2001 .

[84]  Adam D. Bull,et al.  Convergence Rates of Efficient Global Optimization Algorithms , 2011, J. Mach. Learn. Res..

[85]  Benjamin Van Roy,et al.  Learning to Optimize via Posterior Sampling , 2013, Math. Oper. Res..

[86]  Manfred Opper,et al.  A Bayesian approach to on-line learning , 1999 .

[87]  Diego Klabjan,et al.  Improving the Expected Improvement Algorithm , 2017, NIPS.