vrAIn: A Deep Learning Approach Tailoring Computing and Radio Resources in Virtualized RANs

The virtualization of radio access networks (vRAN) is the last milestone in the NFV revolution. However, the complex dependencies between computing and radio resources make vRAN resource control particularly daunting. We present vrAIn, a dynamic resource controller for vRANs based on deep reinforcement learning. First, we use an autoencoder to project high-dimensional context data (traffic and signal quality patterns) into a latent representation. Then, we use a deep deterministic policy gradient (DDPG) algorithm based on an actor-critic neural network structure and a classifier to map (encoded) contexts into resource control decisions. We have implemented vrAIn using an open-source LTE stack over different platforms. Our results show that vrAIn successfully derives appropriate compute and radio control actions irrespective of the platform and context: (i) it provides savings in computational capacity of up to 30% over CPU-unaware methods; (ii) it improves the probability of meeting QoS targets by 25% over static allocation policies using similar CPU resources in average; (iii) upon CPU capacity shortage, it improves throughput performance by 25% over state-of-the-art schemes; and (iv) it performs close to optimal policies resulting from an offline oracle. To the best of our knowledge, this is the first work that thoroughly studies the computational behavior of vRANs, and the first approach to a model-free solution that does not need to assume any particular vRAN platform or system conditions.

[1]  Peter Dayan,et al.  Technical Note: Q-Learning , 2004, Machine Learning.

[2]  Youyong Kong,et al.  Deep Direct Reinforcement Learning for Financial Signal Representation and Trading , 2017, IEEE Transactions on Neural Networks and Learning Systems.

[3]  Fei Wang,et al.  Deep learning for healthcare: review, opportunities and challenges , 2018, Briefings Bioinform..

[4]  George Iosifidis,et al.  FluidRAN: Optimized vRAN/MEC Orchestration , 2018, IEEE INFOCOM 2018 - IEEE Conference on Computer Communications.

[5]  Andres Garcia-Saavedra,et al.  Overbooking network slices through yield-driven end-to-end orchestration , 2018, CoNEXT.

[6]  Liang Tang,et al.  Automatic ad format selection via contextual bandits , 2013, CIKM.

[7]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[8]  Filip De Turck,et al.  Network Function Virtualization: State-of-the-Art and Research Challenges , 2015, IEEE Communications Surveys & Tutorials.

[9]  Chun Yeow Yeoh,et al.  Performance study of LTE experimental testbed using OpenAirInterface , 2016, 2016 18th International Conference on Advanced Communication Technology (ICACT).

[10]  Ambuj Tewari,et al.  From Ads to Interventions: Contextual Bandits in Mobile Health , 2017, Mobile Health - Sensors, Analytic Methods, and Applications.

[11]  Vikram Srinivasan,et al.  CloudIQ: a framework for processing base stations in a data center , 2012, Mobicom '12.

[12]  Andres Garcia-Saavedra,et al.  WizHaul: On the Centralization Degree of Cloud RAN Next Generation Fronthaul , 2018, IEEE Transactions on Mobile Computing.

[13]  Arafat J. Al-Dweik,et al.  QoS-Aware Power-Efficient Scheduler for LTE Uplink , 2014, IEEE Transactions on Mobile Computing.

[14]  Giuseppe Piro,et al.  Downlink Packet Scheduling in LTE Cellular Networks: Key Design Issues and a Survey , 2013, IEEE Communications Surveys & Tutorials.

[15]  Ke Wang,et al.  Computing aware scheduling in mobile edge computing system , 2019, Wirel. Networks.

[16]  M. S. Alam,et al.  Downlink SNR to CQI Mapping for Different Multiple Antenna Techniques in LTE , 2012 .

[17]  Gerhard Fettweis,et al.  Benefits and challenges of virtualization in 5G radio access networks , 2015, IEEE Communications Magazine.

[18]  Kang G. Shin,et al.  RT-OPEX: Flexible Scheduling for Cloud-RAN Processing , 2016, CoNEXT.

[19]  Matthew C. Valenti,et al.  The Complexity–Rate Tradeoff of Centralized Radio Access Networks , 2015, IEEE Transactions on Wireless Communications.

[20]  Paul Turner,et al.  CPU bandwidth control for CFS , 2010 .

[21]  Xiao Zhang,et al.  Autonomous navigation of UAV in large-scale unknown complex environment with deep reinforcement learning , 2017, 2017 IEEE Global Conference on Signal and Information Processing (GlobalSIP).

[22]  Min Sheng,et al.  Max–Min Energy-Efficient Power Allocation in Interference-Limited Wireless Networks , 2015, IEEE Transactions on Vehicular Technology.

[23]  Vincent W. S. Wong,et al.  A Dynamic Resource Sharing Mechanism for Cloud Radio Access Networks , 2016, IEEE Transactions on Wireless Communications.

[24]  Eduardo F. Morales,et al.  An Introduction to Reinforcement Learning , 2011 .

[25]  Marco Fiore,et al.  How Should I Slice My Network?: A Multi-Service Empirical Evaluation of Resource Sharing Efficiency , 2018, MobiCom.

[26]  Mung Chiang,et al.  Multiresource Allocation: Fairness–Efficiency Tradeoffs in a Unifying Framework , 2012, IEEE/ACM Transactions on Networking.

[27]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[28]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[29]  Matthew C. Valenti,et al.  Computationally Aware Sum-Rate Optimal Scheduling for Centralized Radio Access Networks , 2014, 2015 IEEE Global Communications Conference (GLOBECOM).

[30]  Dario Pompili,et al.  Understanding the Computational Requirements of Virtualized Baseband Units Using a Programmable Cloud Radio Access Network Testbed , 2017, 2017 IEEE International Conference on Autonomic Computing (ICAC).

[31]  Jan Peters,et al.  Reinforcement learning in robotics: A survey , 2013, Int. J. Robotics Res..

[32]  Ingrid Moerman,et al.  Cellular Access Multi-Tenancy through Small Cell Virtualization and Common RF Front-End Sharing , 2017, WiNTECH@MobiCom.

[33]  Prostitution In Nevada,et al.  ANNALS of the Association of American Geographers , 1974 .

[34]  A. Stewart Fotheringham,et al.  Principal Component Analysis on Spatial Data: An Overview , 2013 .

[35]  Guy Lever,et al.  Deterministic Policy Gradient Algorithms , 2014, ICML.

[36]  Sergey Levine,et al.  Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[37]  Peter Dayan,et al.  Q-learning , 1992, Machine Learning.

[38]  Song Guo,et al.  Joint Resource Allocation for Max-Min Throughput in Multicell Networks , 2014, IEEE Transactions on Vehicular Technology.

[39]  Demis Hassabis,et al.  Mastering the game of Go with deep neural networks and tree search , 2016, Nature.

[40]  Gerhard Fettweis,et al.  Are Heterogeneous Cloud-Based Radio Access Networks Cost Effective? , 2015, IEEE Journal on Selected Areas in Communications.

[41]  Anas Showk,et al.  Performance analysis of LTE protocol processing on an ARM based mobile platform , 2009, 2009 International Symposium on System-on-Chip.

[42]  Scott Shenker,et al.  PRAN: Programmable Radio Access Networks , 2014, HotNets.

[43]  Peter Rost,et al.  CARES: Computation-Aware Scheduling in Virtualized Radio Access Networks , 2018, IEEE Transactions on Wireless Communications.

[44]  Yuval Tassa,et al.  Continuous control with deep reinforcement learning , 2015, ICLR.

[45]  George Iosifidis,et al.  Joint Optimization of Edge Computing Architectures and Radio Access Networks , 2018, IEEE Journal on Selected Areas in Communications.

[46]  Michael S. Berger,et al.  Cloud RAN for Mobile Networks—A Technology Overview , 2015, IEEE Communications Surveys & Tutorials.

[47]  Cristina Cano,et al.  srsLTE: an open-source platform for LTE evolution and experimentation , 2016, WiNTECH@MobiCom.