Reinforcement Learning for Slicing in a 5G Flexible RAN

Network slicing enables an infrastructure provider (InP) to support heterogeneous 5G services over a common platform (i.e., by creating a customized slice for each service). Once in operation, slices can be dynamically scaled up/down to match the variation of their service requirements. An InP generates revenue by accepting a slice request. If a slice cannot be scaled up when required, an InP has to also pay a penalty (proportional to the level of service degradation). It becomes then crucial for an InP to decide which slice requests should be accepted/rejected in order to increase its net profit. This paper presents a slice admission strategy based on reinforcement learning (RL) in the presence of services with different priorities. The use case considered is a 5G flexible radio access network (RAN), where slices of different mobile service providers are virtualized over the same RAN infrastructure. The proposed policy learns which are the services with the potential to bring high profit (i.e., high revenue with low degradation penalty), and hence should be accepted. The performance of the RL-based admission policy is compared against two deterministic heuristics. Results show that in the considered scenario, the proposed strategy outperforms the benchmark heuristics by at least 23%. Moreover, this paper shows how the policy is able to adapt to different conditions in terms of 1) slice degradation penalty versus slice revenue factors, and 2) proportion of high versus low priority services.

[1]  Andres Garcia-Saavedra,et al.  OVNES: Demonstrating 5G network slicing overbooking on real deployments , 2018, IEEE INFOCOM 2018 - IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS).

[2]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[3]  Di Feng,et al.  Rational Impatience Admission Control in 5G-sliced Networks: Shall I Bide my Slice Opportunity? , 2018 .

[4]  Marco Gramaglia,et al.  Mobile traffic forecasting for maximizing 5G network slicing resource utilization , 2017, IEEE INFOCOM 2017 - IEEE Conference on Computer Communications.

[5]  Lena Wosinska,et al.  A Slice Admission Policy Based on Reinforcement Learning for a 5G Flexible RAN , 2018, 2018 European Conference on Optical Communication (ECOC).

[6]  Raul Muñoz,et al.  Optical Networks Virtualization and Slicing in the 5G era , 2018, 2018 Optical Fiber Communications Conference and Exposition (OFC).

[7]  Chunlei Xu,et al.  A Deep-reinforcement Learning Approach for SDN Routing Optimization , 2020, CSAE.

[8]  Roberto Proietti,et al.  Deep-RMSA: A Deep-Reinforcement-Learning Routing, Modulation and Spectrum Assignment Agent for Elastic Optical Networks , 2018, 2018 Optical Fiber Communications Conference and Exposition (OFC).

[9]  Jing Wang,et al.  A deep reinforcement learning based framework for power-efficient resource allocation in cloud RANs , 2017, 2017 IEEE International Conference on Communications (ICC).

[10]  Lena Wosinska,et al.  Dynamic slicing approach for multi-tenant 5G transport networks [invited] , 2018, IEEE/OSA Journal of Optical Communications and Networking.

[11]  Srikanth Kandula,et al.  Resource Management with Deep Reinforcement Learning , 2016, HotNets.

[12]  Dimitri Ktenas,et al.  QoS-Driven Scheduling in 5G Radio Access Networks - A Reinforcement Learning Approach , 2017, GLOBECOM 2017 - 2017 IEEE Global Communications Conference.

[13]  Xianfu Chen,et al.  Deep Reinforcement Learning for Network Slicing , 2018, ArXiv.

[14]  Vincenzo Sciancalepore,et al.  Slice as a Service (SlaaS) Optimal IoT Slice Resources Orchestration , 2017, GLOBECOM 2017 - 2017 IEEE Global Communications Conference.

[15]  Aric Hagberg,et al.  Exploring Network Structure, Dynamics, and Function using NetworkX , 2008, Proceedings of the Python in Science Conference.

[16]  Di Feng,et al.  A Utility-Driven Multi-Queue Admission Control Solution for Network Slicing , 2019, IEEE INFOCOM 2019 - IEEE Conference on Computer Communications.