Toward Using Reinforcement Learning for Trigger Selection in Network Slice Mobility

Recent 5G trials have demonstrated the usefulness of the Network Slicing concept that delivers customizable services to new and under-serviced industry sectors. However, user mobility’s impact on the optimal resource allocation within and between slices deserves more attention. Slices and their dedicated resources should be offered where the services are to be consumed to minimize network latency and associated overheads and costs. Different mobility patterns lead to different resource re-allocation triggers, leading eventually to slice mobility when enough resources are to be migrated. The selection of the proper triggers for resource re-allocation and related slice mobility patterns is challenging due to triggers’ multiplicity and overlapping nature. In this paper, we investigate the applicability of two Deep Reinforcement Learning based algorithms for allowing a fine-grained selection of mobility triggers that may instantiate slice and resource mobility actions. While the first proposed algorithm relies on a value-based learning method, the second one exploits a hybrid approach to optimize the action selection process. We present an enhanced ETSI Network Function Virtualization edge computing architecture that incorporates the studied mechanisms to implement service and slice migration. We evaluate the proposed methods’ efficiency in a simulated environment and compare their performance in terms of training stability, learning time, and scalability. Finally, we identify and quantify the applicability aspects of the respective approaches.

[1]  Zarita Zainuddin,et al.  Function approximation using artificial neural networks , 2007 .

[2]  Jose Ordonez-Lucena,et al.  Network Slicing for 5G with SDN/NFV: Concepts, Architectures, and Challenges , 2017, IEEE Communications Magazine.

[3]  Richard S. Sutton,et al.  Learning to predict by the methods of temporal differences , 1988, Machine Learning.

[4]  Tarik Taleb,et al.  On Multi-Domain Network Slicing Orchestration Architecture and Federated Resource Control , 2019, IEEE Network.

[5]  Sergey Levine,et al.  Trust Region Policy Optimization , 2015, ICML.

[6]  Tarik Taleb,et al.  MIRA!: An SDN-Based Framework for Cross-Domain Fast Migration of Ultra-Low Latency 5G Services , 2018, 2018 IEEE Global Communications Conference (GLOBECOM).

[7]  Alec Radford,et al.  Proximal Policy Optimization Algorithms , 2017, ArXiv.

[8]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[9]  Wojciech Zaremba,et al.  OpenAI Gym , 2016, ArXiv.

[10]  Tarik Taleb,et al.  Fast Service Migration in 5G Trends and Scenarios , 2020, IEEE Network.

[11]  Yue Wang,et al.  Artificial Intelligence for Elastic Management and Orchestration of 5G Networks , 2019, IEEE Wireless Communications.

[12]  Enda Barrett,et al.  A reinforcement learning approach for dynamic selection of virtual machines in cloud data centres , 2016, 2016 Sixth International Conference on Innovative Computing Technology (INTECH).

[13]  Rajkumar Buyya,et al.  CloudSim: a toolkit for modeling and simulation of cloud computing environments and evaluation of resource provisioning algorithms , 2011, Softw. Pract. Exp..

[14]  Tarik Taleb,et al.  PERMIT: Network Slicing for Personalized 5G Mobile Telecommunications , 2017, IEEE Communications Magazine.

[15]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[16]  Tarik Taleb,et al.  Network Slice Mobility in Next Generation Mobile Systems: Challenges and Potential Solutions , 2020, IEEE Network.

[17]  Pavlin Radoslavov,et al.  ONOS: towards an open, distributed SDN OS , 2014, HotSDN.

[18]  J. Henseler Back Propagation , 1995, Artificial Neural Networks.

[19]  Pasi Liljeberg,et al.  LiRCUP: Linear Regression Based CPU Usage Prediction Algorithm for Live Migration of Virtual Machines in Data Centers , 2013, 2013 39th Euromicro Conference on Software Engineering and Advanced Applications.

[20]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[21]  Hiroaki Harai,et al.  Reinforcement learning based dynamic resource migration for virtual networks , 2017, 2017 IFIP/IEEE Symposium on Integrated Network and Service Management (IM).

[22]  Andrew W. Moore,et al.  Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..

[23]  Matti Latva-aho,et al.  Key drivers and research challenges for 6G ubiquitous wireless intelligence , 2019 .

[24]  David A. Patterson,et al.  In-datacenter performance analysis of a tensor processing unit , 2017, 2017 ACM/IEEE 44th Annual International Symposium on Computer Architecture (ISCA).

[25]  Peter Dayan,et al.  Q-learning , 1992, Machine Learning.

[26]  Tarik Taleb,et al.  Towards Studying Service Function Chain Migration Patterns in 5G Networks and Beyond , 2019, 2019 IEEE Global Communications Conference (GLOBECOM).

[27]  David Bernstein,et al.  Containers and Cloud: From LXC to Docker to Kubernetes , 2014, IEEE Cloud Computing.

[28]  Nei Kato,et al.  Machine Learning Meets Computation and Communication Control in Evolving Edge and Cloud: Challenges and Future Perspective , 2020, IEEE Communications Surveys & Tutorials.

[29]  Tarik Taleb,et al.  Network Slicing and Softwarization: A Survey on Principles, Enabling Technologies, and Solutions , 2018, IEEE Communications Surveys & Tutorials.

[30]  Xin Wang,et al.  Machine Learning for Networking: Workflow, Advances and Opportunities , 2017, IEEE Network.

[31]  Vaishnavi Ravi,et al.  Reinforcement Learning Based Service Provisioning for a Greener Cloud , 2014, 2014 3rd International Conference on Eco-friendly Computing and Communication Systems.

[32]  Yishay Mansour,et al.  Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.

[33]  Tarik Taleb,et al.  Dynamic Resource Provisioning of a Scalable E2E Network Slicing Orchestration System , 2020, IEEE Transactions on Mobile Computing.

[34]  White Paper 5G Evolution and 6G , 2020 .

[35]  Mehdi Bennis,et al.  Wireless Network Intelligence at the Edge , 2018, Proceedings of the IEEE.

[36]  Christos V. Verikoukis,et al.  Software Defined Network Service Chaining for OTT Service Providers in 5G Networks , 2017, IEEE Communications Magazine.

[37]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[38]  Tarik Taleb,et al.  Towards a Fast Service Migration in 5G , 2018, 2018 IEEE Conference on Standards for Communications and Networking (CSCN).

[39]  Tarik Taleb,et al.  On Multi-Access Edge Computing: A Survey of the Emerging 5G Network Edge Cloud Architecture and Orchestration , 2017, IEEE Communications Surveys & Tutorials.

[40]  Geoffrey E. Hinton,et al.  Replicated Softmax: an Undirected Topic Model , 2009, NIPS.

[41]  Enda Barrett,et al.  A Predictive Anti-Correlated Virtual Machine Placement Algorithm for Green Cloud Computing , 2018, 2018 IEEE/ACM 11th International Conference on Utility and Cloud Computing (UCC).

[42]  Tarik Taleb,et al.  A Novel QoS Framework for Network Slicing in 5G and Beyond Networks Based on SDN and NFV , 2020, IEEE Network.

[43]  Gulshan Kumar,et al.  A Survey of Deep Learning and Its Applications: A New Paradigm to Machine Learning , 2019, Archives of Computational Methods in Engineering.

[44]  Andrew L. Maas Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .

[45]  Tarik Taleb,et al.  End-to-end Network Slicing for 5G Mobile Networks , 2017, J. Inf. Process..

[46]  Nando de Freitas,et al.  Sample Efficient Actor-Critic with Experience Replay , 2016, ICLR.

[47]  Marc Peter Deisenroth,et al.  Deep Reinforcement Learning: A Brief Survey , 2017, IEEE Signal Processing Magazine.

[48]  Eduardo F. Morales,et al.  An Introduction to Reinforcement Learning , 2011 .

[49]  Ying-Chang Liang,et al.  Applications of Deep Reinforcement Learning in Communications and Networking: A Survey , 2018, IEEE Communications Surveys & Tutorials.

[50]  Alex Graves,et al.  Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.

[51]  Helmut Hlavacs,et al.  An Intelligent and Adaptive Threshold-Based Schema for Energy and Performance Efficient Dynamic VM Consolidation , 2013, EE-LSDS.

[52]  Tarik Taleb,et al.  Network Slicing-Based Customization of 5G Mobile Services , 2019, IEEE Network.

[53]  Richard Evans,et al.  Deep Reinforcement Learning in Large Discrete Action Spaces , 2015, 1512.07679.

[54]  Tom Schaul,et al.  Prioritized Experience Replay , 2015, ICLR.

[55]  Wansu Lim,et al.  Machine Learning for 5G/B5G Mobile and Wireless Communications: Potential, Limitations, and Future Directions , 2019, IEEE Access.

[56]  Natalia Gimelshein,et al.  PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.