Optimal sampling of genetic algorithms on polynomial regression

This paper investigates the utility of sampling as an evaluation-relaxation technique in genetic algorithms (GAs). In many real-world applications, sampling can be used to generate a less accurate, but computationally inexpensive fitness evaluator to speed GAs up. This paper focuses on the problem of polynomial regression as an example of problems with positive dependency among genes. Via statistical analysis of the noise introduced by sampling, this paper develops facet-wise models for the optimal sampling size, and these models are empirically verified. The results show that when the population is sized properly, small sampling sizes are preferred for most applications. When a fixed population size is adopted, which is usually the case in real-world applications, an optimal sampling size exists. If the sampling size is too small, the sampling noise increases, and GAs would perform poorly because of an insufficiently large population. If the sampling size is too large, the GA would spend too much time in fitness calculation and cannot perform well either within limited run duration.

[1]  Thomas Bäck,et al.  Evolution Strategies on Noisy Functions: How to Improve Convergence Properties , 1994, PPSN.

[2]  John J. Grefenstette,et al.  Genetic Search with Approximate Function Evaluation , 1985, ICGA.

[3]  Philippe Giguere Population Sizing for Optimum Sampling with Genetic Algorithms: A Case Study of the Onemax Problem , 1998 .

[4]  D. E. Goldberg,et al.  Simple Genetic Algorithms and the Minimal, Deceptive Problem , 1987 .

[5]  Jacob Cohen,et al.  Applied multiple regression/correlation analysis for the behavioral sciences , 1979 .

[6]  Brad L. Miller,et al.  Noise, sampling, and efficient genetic algorthms , 1997 .

[7]  Kalyanmoy Deb,et al.  Analyzing Deception in Trap Functions , 1992, FOGA.

[8]  David E. Goldberg,et al.  The Gambler's Ruin Problem, Genetic Algorithms, and the Sizing of Populations , 1999, Evolutionary Computation.

[9]  David E. Goldberg,et al.  Optimal Sampling and Speed-Up for Genetic Algorithms on the Sampled OneMax Problem , 2003, GECCO.

[10]  Dirk Thierens,et al.  Convergence Models of Genetic Algorithm Selection Schemes , 1994, PPSN.

[11]  Heinz Mühlenbein,et al.  Predictive Models for the Breeder Genetic Algorithm I. Continuous Parameter Optimization , 1993, Evolutionary Computation.

[12]  David E. Goldberg,et al.  Genetic Algorithms, Selection Schemes, and the Varying Effects of Noise , 1996, Evolutionary Computation.

[13]  Benjamin W. Wah,et al.  Scheduling of Genetic Algorithms in a Noisy Environment , 1994, Evolutionary Computation.

[14]  John R. Koza,et al.  Genetic programming - on the programming of computers by means of natural selection , 1993, Complex adaptive systems.

[15]  Hans-Georg Beyer,et al.  Toward a Theory of Evolution Strategies: Some Asymptotical Results from the (1,+ )-Theory , 1993, Evolutionary Computation.

[16]  David E. Goldberg,et al.  Optimal Sampling For Genetic Algorithms , 1996 .

[17]  Zbigniew Michalewicz,et al.  Evolutionary Computation 1 , 2018 .

[18]  Kalyanmoy Deb,et al.  Genetic Algorithms, Noise, and the Sizing of Populations , 1992, Complex Syst..

[19]  David E. Goldberg,et al.  The Design of Innovation: Lessons from and for Competent Genetic Algorithms , 2002 .

[20]  David E. Goldberg,et al.  Genetic Algorithms in Search Optimization and Machine Learning , 1988 .

[21]  D. E. Goldberg,et al.  Genetic Algorithms in Search , 1989 .

[22]  David E. Goldberg,et al.  The Race, the Hurdle, and the Sweet Spot , 1998 .

[23]  Lothar Thiele,et al.  A Mathematical Analysis of Tournament Selection , 1995, ICGA.

[24]  David E. Goldberg,et al.  Genetic Algorithms, Efficiency Enhancement, And Deciding Well With Differing Fitness Variances , 2002, GECCO.

[25]  Magnus Rattray,et al.  Noisy Fitness Evaluation in Genetic Algorithms and the Dynamics of Learning , 1996, FOGA.