Improving Search Algorithms by Using Intelligent Coordinates

We consider algorithms that maximize a global function G in a distributed manner, using a different adaptive computational agent to set each variable of the underlying space. Each agent eta is self-interested; it sets its variable to maximize its own function g(eta). Three factors govern such a distributed algorithm's performance, related to exploration/exploitation, game theory, and machine learning. We demonstrate how to exploit all three factors by modifying a search algorithm's exploration stage: rather than random exploration, each coordinate of the search space is now controlled by a separate machine-learning-based "player" engaged in a noncooperative game. Experiments demonstrate that this modification improves simulated annealing (SA) by up to an order of magnitude for bin packing and for a model of an economic process run over an underlying network. These experiments also reveal interesting small-world phenomena.

[1]  C. D. Gelatt,et al.  Optimization by Simulated Annealing , 1983, Science.

[2]  Donald Geman,et al.  Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images , 1984, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[3]  Editors , 1986, Brain Research Bulletin.

[4]  Stuart German,et al.  Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images , 1988 .

[5]  Bernardo A. Huberman,et al.  The ecology of computation , 1988, Digest of Papers. COMPCON Spring 89. Thirty-Fourth IEEE Computer Society International Conference: Intellectual Leverage.

[6]  Drew Fudenberg,et al.  Game theory (3. pr.) , 1991 .

[7]  Andrew G. Barto,et al.  Improving Elevator Performance Using Reinforcement Learning , 1995, NIPS.

[8]  Zbigniew Michalewicz,et al.  Handbook of Evolutionary Computation , 1997 .

[9]  M. Marsili,et al.  A Prototype Model of Stock Exchange , 1997, cond-mat/9709118.

[10]  Jeffrey Horn,et al.  Handbook of evolutionary computation , 1997 .

[11]  Physics Department,et al.  Adaptive Competition, Market Efficiency, Phase Transitions and Spin-Glasses , 1997 .

[12]  D. Saad Europhysics Letters , 1997 .

[13]  Yicheng Zhang Modeling Market Mechanism with Evolutionary Games , 1998, cond-mat/9803308.

[14]  Duncan J. Watts,et al.  Collective dynamics of ‘small-world’ networks , 1998, Nature.

[15]  P. Pardalos,et al.  Handbook of Combinatorial Optimization , 1998 .

[16]  E. Coffman,et al.  Bin Packing Approximation Algorithms: Combinatorial Analysis , 1999, Handbook of Combinatorial Optimization.

[17]  Kagan Tumer,et al.  Collective Intelligence for Control of Distributed Dynamical Systems , 1999, ArXiv.

[18]  David B. Fogel,et al.  Evolution, neural networks, games, and intelligence , 1999, Proc. IEEE.

[19]  Stroud,et al.  Exact results and scaling properties of small-world networks , 1999, Physical review. E, Statistical physics, plasmas, fluids, and related interdisciplinary topics.

[20]  M. Newman,et al.  Mean-field solution of the small-world network model. , 1999, Physical review letters.

[21]  Kagan Tumer,et al.  Collective Intelligence and Braess' Paradox , 2000, AAAI/IAAI.

[22]  Laurent Keller,et al.  Ant-like task allocation and recruitment in cooperative robots , 2000, Nature.

[23]  G. Theraulaz,et al.  Inspiration for optimization from social insect behaviour , 2000, Nature.

[24]  Noam Nisan,et al.  Algorithmic Mechanism Design , 2001, Games Econ. Behav..

[25]  Kagan Tumer,et al.  Optimal Payoff Functions for Members of Collectives , 2001, Adv. Complex Syst..

[26]  Damien Challet,et al.  Optimal combinations of imperfect objects. , 2002, Physical review letters.

[27]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.