A Scalable Deep Reinforcement Learning Approach for Traffic Engineering Based on Link Control

As modern communication networks are growing more complicated and dynamic, designing a good Traffic Engineering (TE) policy becomes difficult due to the complexity of solving the optimal traffic scheduling problem. Deep Reinforcement Learning (DRL) provides us with a chance to design a model-free TE scheme through machine learning. However, existing DRL-based TE solutions cannot be applied to large networks. In this article, we propose to combine the control theory and DRL to design a TE scheme. Our proposed scheme ScaleDRL employs the idea from the pinning control theory to select a subset of links in the network and name them critical links. Based on the traffic distribution information, we use a DRL algorithm to dynamically adjust the link weights for the critical links. Through a weighted shortest path algorithm, the forwarding paths of the flows can be dynamically adjusted. The packet-level simulation shows that ScaleDRL reduces the average end-to-end transmission delay by up to 39% compared to the state-of-the-art in different network topologies.

[1]  H. Jonathan Chao,et al.  SmartEntry: Mitigating Routing Update Overhead with Reinforcement Learning for Traffic Engineering , 2020, NetAI@SIGCOMM.

[2]  Hamed Haddadi,et al.  Deep Learning in Mobile and Wireless Networking: A Survey , 2018, IEEE Communications Surveys & Tutorials.

[3]  Min Chen,et al.  TIDE: Time-relevant deep reinforcement learning for routing optimization , 2019, Future Gener. Comput. Syst..

[4]  Ian F. Akyildiz,et al.  QoS-Aware Adaptive Routing in Multi-layer Hierarchical Software Defined Networks: A Reinforcement Learning Approach , 2016, 2016 IEEE International Conference on Services Computing (SCC).

[5]  Hong Xu,et al.  Node-Constrained Traffic Engineering: Theory and Applications , 2019, IEEE/ACM Transactions on Networking.

[6]  Albert-László Barabási,et al.  Controllability of complex networks , 2011, Nature.

[7]  Hans-Werner Braun,et al.  The National Science Foundation Network , 1992 .

[8]  Elman Mansimov,et al.  Scalable trust-region method for deep reinforcement learning using Kronecker-factored approximation , 2017, NIPS.

[9]  Shane Legg,et al.  Noisy Networks for Exploration , 2017, ICLR.

[10]  Chi Harold Liu,et al.  Experience-driven Networking: A Deep Reinforcement Learning based Approach , 2018, IEEE INFOCOM 2018 - IEEE Conference on Computer Communications.

[11]  Junjie Zhang,et al.  CFR-RL: Traffic Engineering With Reinforcement Learning in SDN , 2020, IEEE Journal on Selected Areas in Communications.

[12]  Ibrahim Matta,et al.  BRITE: an approach to universal topology generation , 2001, MASCOTS 2001, Proceedings Ninth International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems.

[13]  Yuan Zuo,et al.  Learning-based network path planning for traffic engineering , 2019, Future Gener. Comput. Syst..

[14]  M. Köppen,et al.  The Curse of Dimensionality , 2010 .

[15]  Konstantina Papagiannaki,et al.  Structural analysis of network traffic flows , 2004, SIGMETRICS '04/Performance '04.

[16]  Giuseppe Aceto,et al.  Toward effective mobile encrypted traffic classification through deep learning , 2020, Neurocomputing.

[17]  Alex Graves,et al.  Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.