Thèse d'habilitation à diriger des recherches "Analysis of Comparison-based Stochastic Continuous Black-Box Optimization Algorithms". (Thèse d'habilitation à diriger des recherches "Analysis of Comparison-based Stochastic Continuous Black-Box Optimization Algorithms")

This manuscript presents a large part of my research since the end of my PhD. Most of my work is related to numerical (also referred to as continuous) optimization, at the exception of one contribution done during my postdoc in Zurich introducing a new stochastic algorithm to simulate chemical or biochemical systems [23]. The optimization algorithms at the core of my work are adaptive derivative-free stochastic (or randomized) optimization methods. The algorithms are tailored to tackle dificult numerical optimization problems in a so-called black-box context where the objective function to be optimized is seen as a black-box. For a given input solution, the black-box returns solely the objective function value but no gradient or higher order derivatives are assumed. The optimization algorithm can use the information returned by the black-box, i.e. the history of function values associated to the queried search points, but no other knowledge that could be within the black-box (parameters describing the class of functions the function belongs to, ...). This black-box context is very natural in industrial settings where the function to be optimized can be given by an executable file for which the source code is not provided. It is also natural in situations where the function is given by a large simulation code from which it is hard to extract any useful information for the optimization. This context is also called derivative-free optimization (DFO) in the mathematical optimization community. Well-known DFO methods are the Nelder-Mead algorithm [79, 77], pattern search methods [54, 90, 6] or more recently the NEW Unconstraint Optimization Algorithm (NEWUOA) developed by Powell [82, 81]. In this context, I have been focusing on DFO methods in the literal sense. However the methods my research is centered on have a large stochastic component and originate from the community of bio-inspired algorithms mainly composed of computer scientists and engineers. The methods were introduced at the end of the 70's. A parallel with Darwin's theory of the evolution of species based on blind variation and natural selection was recognized and served as source of inspiration for those methods. Nowadays this field of bio-inspired methods is referred to as evolutionary computation (EC) and a generic term for the methods is evolutionary algorithms. The probably most famous examples of bio-inspired methods are genetic algorithms (GAs). However today GAs are known to be not competitive for numerical optimization. Evolution Strategies (ES) introduced in the end of the 70's [83] have emerged as the main sub-branch of EC devoted to continuous optimization. One important feature of ES is that they are comparison-based algorithms. The present most advanced ES algorithm, the Covariance Matrix Adaptation Evolution Strategy (CMA-ES) [50] is a variable metric method recognized as the state-of-the-art method for stochastic numerical optimization. It is used in many applications in industry and academy. Because of historical reasons, the developments and work on Evolution Strategies are mainly carried out in the EC field where practice and effectiveness is definitely as (or more) important as having a theorem proven about an algorithm. However ES algorithms are simply adaptive stochastic iterative methods and they need to be studied from a mathematical perspective as well as any other iterative method in optimization or other domain in order to understand the methods better and convince a broader class of people about their soundness. Questions like their convergence and speed of convergence central in optimization need to be addressed. My research is encompassed within this general context: I am particularly interested by the mathematical aspects of adaptive stochastic methods like ES (and of course CMA-ES) or more generally adaptive stochastic optimization algorithms. Evolution strategies have this attractive facet that while introduced in the bio-inspired and engineering context, they turn out to be methods with deep theoretical foundations related to invariance, information geometry, stochastic approximation and strongly connected to Markov chain Monte Carlo (MCMC) algorithms. Those foundations and connections are relatively new and to a small (for some topics) or large (for others) extent partly related to some of my contributions. They will be explained within the manuscript. I particularly care that the theory I am working on relates to practical algorithms or has an impact on (new) algorithm designs. I attempt to illustrate this within the manuscript. While optimization is the central theme of my research, I have been tackling various aspect of optimization. Although most of my work is devoted to single-objective optimization, I have also been working on multi-objective optimization where the goal is to optimize simultaneously several conflicting objectives and where instead of a single solution, a set of solutions, the so-called Pareto set composed of the best compromises is searched. In the field of single-objective optimization, I have been tackling diverse contexts like noisy optimization where for a given point in a search space we do not observe one deterministic value but a distribution of possible function values, large-scale optimization where one is interested in tackling problems of the order of 104 (medium large-scale) to 106 variables (large-scale) and to a smaller extent constrained optimization. In addition to investigating theoretical questions, I have been also working on designing new algorithms that calls for theory complemented with numerical simulations. Last I have tackled some applications mainly in the context of the PhD of Mohamed Jebalia with an application in chromatography and of the PhD of Zyed Bouzarkouna (PhD financed by the French Institute for petrol) on the placement of oil wells. Furthermore, a non neglect-able part of my research those past years has been devoted to benchmarking of algorithms. Benchmarking complements theory as it is difficult to assess theoretically the performance of algorithms on all typical functions one is interested. The main motivation has then been to improve the standards on how benchmarking is done. Those contributions were done along with the development of the Comparing COntinuous Optimizers platform (COCO). My work is articulated around three main complementary axis, namely theory / algorithm design and applications. An overview of the contributions presented within this habilitation organized along those axes is given in Figure 3.1.

[1]  Lothar Thiele,et al.  Multiobjective Optimization Using Evolutionary Algorithms - A Comparative Case Study , 1998, PPSN.

[2]  M. Powell Developments of NEWUOA for unconstrained minimization without derivatives , 2007 .

[3]  K. Steiglitz,et al.  Adaptive step size random search , 1968 .

[4]  Xin Yao,et al.  Evolutionary programming made faster , 1999, IEEE Trans. Evol. Comput..

[5]  Virginia Torczon,et al.  On the Convergence of Pattern Search Algorithms , 1997, SIAM J. Optim..

[6]  K. I. M. McKinnon,et al.  Convergence of the Nelder-Mead Simplex Method to a Nonstationary Point , 1998, SIAM J. Optim..

[7]  Günter Rudolph,et al.  Local convergence rates of simple evolutionary algorithms with Cauchy mutations , 1997, IEEE Trans. Evol. Comput..

[8]  M. Powell The NEWUOA software for unconstrained optimization without derivatives , 2006 .

[9]  John A. Nelder,et al.  A Simplex Method for Function Minimization , 1965, Comput. J..

[10]  I. Loshchilov,et al.  Achieving optimization invariance w.r.t. monotonous transformations of the objective function and orthogonal transformations of the representation , 2012, NIPS 2012.

[11]  Ingo Rechenberg,et al.  Evolutionsstrategie : Optimierung technischer Systeme nach Prinzipien der biologischen Evolution , 1973 .

[12]  Richard L. Tweedie,et al.  Markov Chains and Stochastic Stability , 1993, Communications and Control Engineering Series.

[13]  James C. Spall,et al.  Adaptive stochastic approximation by the simultaneous perturbation method , 2000, IEEE Trans. Autom. Control..

[14]  Rainer Storn,et al.  Differential Evolution – A Simple and Efficient Heuristic for global Optimization over Continuous Spaces , 1997, J. Glob. Optim..