Consensus-based optimization methods converge globally in mean-field law

In this paper we study consensus-based optimization (CBO), which is a multi-agent metaheuristic derivative-free optimization method that can globally minimize nonconvex nonsmooth functions and is amenable to theoretical analysis. Based on an experimentally supported intuition that CBO always performs a gradient descent of the squared Euclidean distance to the global minimizer, we derive a novel technique for proving the convergence to the global minimizer in mean-field law for a rich class of objective functions. The result unveils internal mechanisms of CBO that are responsible for the success of the method. In particular, we prove that CBO performs a convexification of a very large class of optimization problems as the number of optimizing agents goes to infinity. Furthermore, we improve prior analyses by requiring minimal assumptions about the initialization of the method and by covering objectives that are merely locally Lipschitz continuous. As a by-product of the analysis, we establish a quantitative nonasymptotic Laplace principle, which may be of independent interest.

[1]  W. K. Hastings,et al.  Monte Carlo Sampling Methods Using Markov Chains and Their Applications , 1970 .

[2]  G. Nemhauser,et al.  Wiley‐Interscience Series in Discrete Mathematics and Optimization , 2014 .

[3]  Doheon Kim,et al.  Convergence of a first-order consensus-based global optimization algorithm , 2019, Mathematical Models and Methods in Applied Sciences.

[4]  Mark W. Schmidt,et al.  Linear Convergence of Gradient and Proximal-Gradient Methods Under the Polyak-Łojasiewicz Condition , 2016, ECML/PKDD.

[5]  John H. Holland,et al.  Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence , 1992 .

[6]  Marco Dorigo,et al.  Ant colony optimization theory: A survey , 2005, Theor. Comput. Sci..

[7]  Thomas Bäck,et al.  Evolutionary computation: Toward a new philosophy of machine intelligence , 1997, Complex..

[8]  Shi Jin,et al.  A Consensus-Based Global Optimization Method with Adaptive Momentum Estimation , 2020, Communications in Computational Physics.

[9]  Bruce W. Suter,et al.  From error bounds to the complexity of first-order descent methods for convex functions , 2015, Math. Program..

[10]  Riccardo Poli,et al.  Particle swarm optimization , 1995, Swarm Intelligence.

[11]  Emile H. L. Aarts,et al.  Simulated annealing and Boltzmann machines - a stochastic approach to combinatorial optimization and neural computing , 1990, Wiley-Interscience series in discrete mathematics and optimization.

[12]  M. Fornasier,et al.  Consensus-based optimization on hypersurfaces: Well-posedness and mean-field limit , 2020, Mathematical Models and Methods in Applied Sciences.

[13]  Jos'e A. Carrillo,et al.  An analytical framework for a consensus-based global optimization method , 2016, 1602.00220.

[14]  A. Guillin,et al.  On the rate of convergence in Wasserstein distance of the empirical measure , 2013, 1312.2128.

[15]  Yurii Nesterov,et al.  Linear convergence of first order methods for non-strongly convex optimization , 2015, Math. Program..

[16]  Amir Dembo,et al.  Large Deviations Techniques and Applications , 1998 .

[17]  John A. Nelder,et al.  A Simplex Method for Function Minimization , 1965, Comput. J..

[18]  Christian Blum,et al.  Metaheuristics in combinatorial optimization: Overview and conceptual comparison , 2003, CSUR.

[19]  Dan Boneh,et al.  On genetic algorithms , 1995, COLT '95.

[20]  Seung-Yeal Ha,et al.  A Stochastic Consensus Method for Nonconvex Optimization on the Stiefel Manifold , 2020, 2020 59th IEEE Conference on Decision and Control (CDC).

[21]  Desmond J. Higham,et al.  An Algorithmic Introduction to Numerical Simulation of Stochastic Differential Equations , 2001, SIAM Rev..

[22]  Doheon Kim,et al.  Convergence and error estimates for time-discrete consensus-based optimization algorithms , 2020, Numerische Mathematik.

[23]  Seung-Yeal Ha,et al.  Emergent behaviors of high-dimensional Kuramoto models on Stiefel manifolds , 2021, Autom..

[24]  P. Miller Applied asymptotic analysis , 2006 .

[25]  R. Pinnau,et al.  A consensus-based model for global optimization and its mean-field limit , 2016, 1604.05648.

[26]  Mihai Anitescu,et al.  Degenerate Nonlinear Programming with a Quadratic Growth Condition , 1999, SIAM J. Optim..

[27]  Marie-Therese Wolfram,et al.  Consensus-based global optimization with personal best. , 2020, Mathematical biosciences and engineering : MBE.

[28]  Shi Jin,et al.  A consensus-based global optimization method for high dimensional machine learning problems , 2019 .

[29]  Tianbao Yang,et al.  Adaptive SVRG Methods under Error Bound Conditions with Unknown Growth Parameter , 2017, NIPS.

[30]  L. Ambrosio,et al.  Gradient Flows: In Metric Spaces and in the Space of Probability Measures , 2005 .

[31]  E. Platen An introduction to numerical methods for stochastic differential equations , 1999, Acta Numerica.

[32]  Jeffrey Horn,et al.  Handbook of evolutionary computation , 1997 .

[33]  D. Williams STOCHASTIC DIFFERENTIAL EQUATIONS: THEORY AND APPLICATIONS , 1976 .