Many real-world applications require optimization in dynamic environments where the challenge is to find optima of a time-dependent objective function while tracking them over time. Many evolutionary approaches have been developed to solve Dynamic Optimization Problems (DOPs). However, there is still a need for more efficient methods. Recently, a new interesting trend in dealing with optimization in dynamic environments has emerged toward developing new Reinforcement Learning (RL) algorithms that are expected to give a new breath in DOPs community. In this paper, a new Q-learning RL algorithm is developed to deal with DOPs based on new deifined states and actions that are mainly inspired from Evolutionary Dynamic Optimization (EDO) aiming appropriate exploitation of the strengths of both RL and EDO techniques to handle DOPs. The proposed RL model has been assessed using modified Moving Peaks Benchmark (mMPB) problem. Very competitive results have been obtained and good performance has been achieved compared with other dynamic optimization algorithm.
[1]
Souham Meshoul,et al.
Effective Streaming Evolutionary Feature Selection Using Dynamic Optimization
,
2018,
CIIA.
[2]
Wenbo Xu,et al.
Particle swarm optimization with particles having quantum behavior
,
2004,
Proceedings of the 2004 Congress on Evolutionary Computation (IEEE Cat. No.04TH8753).
[3]
Xin Yao,et al.
What are dynamic optimization problems?
,
2014,
2014 IEEE Congress on Evolutionary Computation (CEC).
[4]
Richard S. Sutton,et al.
Reinforcement Learning: An Introduction
,
1998,
IEEE Trans. Neural Networks.
[5]
Jürgen Branke,et al.
Memory enhanced evolutionary algorithms for changing optimization problems
,
1999,
Proceedings of the 1999 Congress on Evolutionary Computation-CEC99 (Cat. No. 99TH8406).