Power Consumption Optimization Using Gradient Boosting Aided Deep Q-Network in C-RANs

Cloud Radio Access Networks (C-RANs) have the potential to enable growing data traffic in 5G networks. However, with the complex states, resource allocation in C-RANs is time-consuming and computational-expensive, making it challenging to meet the demands of energy efficiency and low latency in real-time wireless applications. In this paper, we propose a gradient boosting decision tree (GBDT)-based deep Q-network (DQN) framework for solving the dynamic resource allocation (DRA) problem in a real-time C-RAN, where the heavy computation to solve SOCP problems is cut down and significant power consumption can be saved. First, we apply the GBDT to the regression task to approximate the solutions of second order cone programming (SOCP) problem formulated from beamforming design which consumes heavy computing resources by traditional algorithms. Then, we design a deep Q-network (DQN), coming from a common deep reinforcement learning, to autonomously generate the robust policy that controls the status of remote radio heads (RRHs) and saves the power consumption in long term. The DQN deploys deep neural networks (DNN) to solve the problem of innumerable states in the real-time C-RAN system and generates the policy by observing the state and the reward engendered by GBDT. The generated policy is error-tolerant considering that the gradient boosting regression may not be strictly subject to the constraints of the original problem. Simulation results validate its advantages in terms of the performance and computational complexity for power consumption saving compared with existing methods.

[1]  Eduardo F. Morales,et al.  An Introduction to Reinforcement Learning , 2011 .

[2]  Tianqi Chen,et al.  XGBoost: A Scalable Tree Boosting System , 2016, KDD.

[3]  Lior Rokach,et al.  Ensemble-based classifiers , 2010, Artificial Intelligence Review.

[4]  Tiejun Lv,et al.  Deep reinforcement learning based computation offloading and resource allocation for MEC , 2018, 2018 IEEE Wireless Communications and Networking Conference (WCNC).

[5]  Wei Yu,et al.  Energy Efficiency of Downlink Transmission Strategies for Cloud Radio Access Networks , 2016, IEEE Journal on Selected Areas in Communications.

[6]  Dirk Wübben,et al.  Cloud technologies for flexible 5G radio access networks , 2014, IEEE Communications Magazine.

[7]  Mugen Peng,et al.  Deep Reinforcement Learning Based Coded Caching Scheme in Fog Radio Access Networks , 2018, 2018 IEEE/CIC International Conference on Communications in China (ICCC Workshops).

[8]  Thar Baker,et al.  PriNergy: a priority-based energy-efficient routing method for IoT systems , 2020, The Journal of Supercomputing.

[9]  J. Friedman Greedy function approximation: A gradient boosting machine. , 2001 .

[10]  Leo Breiman,et al.  Bias, Variance , And Arcing Classifiers , 1996 .

[11]  Sergey D. Andreev,et al.  Cooperative Radio Resource Management in Heterogeneous Cloud Radio Access Networks , 2015, IEEE Access.

[12]  Knud D. Andersen,et al.  The Mosek Interior Point Optimizer for Linear Programming: An Implementation of the Homogeneous Algorithm , 2000 .

[13]  Tao Jiang,et al.  Deep Reinforcement Learning for Mobile Edge Caching: Review, New Features, and Open Issues , 2018, IEEE Network.

[14]  Mahsa Shoaran,et al.  Energy-Efficient Classification for Resource-Constrained Biomedical Applications , 2018, IEEE Journal on Emerging and Selected Topics in Circuits and Systems.

[15]  Kezhi Wang,et al.  Bit-Level Optimized Neural Network for Multi-Antenna Channel Quantization , 2019, IEEE Wireless Communications Letters.

[16]  Hui Tian,et al.  Joint Power and Bandwidth Allocation Algorithm with QoS Support in Heterogeneous Wireless Networks , 2012, IEEE Communications Letters.

[17]  Lisa Turner,et al.  Applications of Second Order Cone Programming , 2012 .

[18]  Young Min Kim,et al.  RL-GAN-Net: A Reinforcement Learning Agent Controlled GAN Network for Real-Time Point Cloud Shape Completion , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[19]  Tao Jiang,et al.  Caching Transient Data for Internet of Things: A Deep Reinforcement Learning Approach , 2019, IEEE Internet of Things Journal.

[20]  Shalabh Bhatnagar,et al.  Efficient Adaptive Resource Provisioning for Cloud Applications using Reinforcement Learning , 2019, 2019 IEEE 4th International Workshops on Foundations and Applications of Self* Systems (FAS*W).

[21]  Peter Henderson,et al.  An Introduction to Deep Reinforcement Learning , 2018, Found. Trends Mach. Learn..

[22]  Jing Wang,et al.  A deep reinforcement learning based framework for power-efficient resource allocation in cloud RANs , 2017, 2017 IEEE International Conference on Communications (ICC).

[23]  Robert Schober,et al.  User Association in 5G Networks: A Survey and an Outlook , 2015, IEEE Communications Surveys & Tutorials.

[24]  R. Polikar,et al.  Ensemble based systems in decision making , 2006, IEEE Circuits and Systems Magazine.

[25]  Ami Wiesel,et al.  Linear precoding via conic optimization for fixed MIMO receivers , 2006, IEEE Transactions on Signal Processing.

[26]  D. Opitz,et al.  Popular Ensemble Methods: An Empirical Study , 1999, J. Artif. Intell. Res..

[27]  Alois Knoll,et al.  Gradient boosting machines, a tutorial , 2013, Front. Neurorobot..

[28]  Jonathan Le Roux,et al.  Deep Unfolding: Model-Based Inspiration of Novel Deep Architectures , 2014, ArXiv.

[29]  Muhammad Ali Imran,et al.  How much energy is needed to run a wireless network? , 2011, IEEE Wireless Communications.

[30]  Yann LeCun,et al.  Learning Fast Approximations of Sparse Coding , 2010, ICML.

[31]  Yuanming Shi,et al.  Group Sparse Beamforming for Green Cloud-RAN , 2013, IEEE Transactions on Wireless Communications.

[32]  Wei Jiang,et al.  Dynamic Reservation and Deep Reinforcement Learning Based Autonomous Resource Slicing for Virtualized Radio Access Networks , 2019, IEEE Access.

[33]  Kezhi Wang,et al.  MIMO Channel Information Feedback Using Deep Recurrent Network , 2018, IEEE Communications Letters.

[34]  Ji Feng,et al.  Deep forest , 2017, IJCAI.

[35]  Jos F. Sturm,et al.  A Matlab toolbox for optimization over symmetric cones , 1999 .

[36]  Geoffrey Ye Li,et al.  Model-Driven Deep Learning for Physical Layer Communications , 2018, IEEE Wireless Communications.

[37]  N. Sidiropoulos,et al.  Learning to Optimize: Training Deep Neural Networks for Interference Management , 2017, IEEE Transactions on Signal Processing.

[38]  Inkyu Lee,et al.  Online Reinforcement Learning of X-Haul Content Delivery Mode in Fog Radio Access Networks , 2019, IEEE Signal Processing Letters.

[39]  Manuel López Martín,et al.  Neural network architecture based on gradient boosting for IoT traffic prediction , 2019, Future Gener. Comput. Syst..

[40]  Michael S. Berger,et al.  Cloud RAN for Mobile Networks—A Technology Overview , 2015, IEEE Communications Surveys & Tutorials.

[41]  Stephen P. Boyd,et al.  ECOS: An SOCP solver for embedded systems , 2013, 2013 European Control Conference (ECC).

[42]  Yun Hee Kim,et al.  Resource Allocation for a Full-Duplex Wireless-Powered Communication Network With Imperfect Self-Interference Cancelation , 2016, IEEE Communications Letters.

[43]  Dirk Van,et al.  Ensemble Methods: Foundations and Algorithms , 2012 .

[44]  Alireza Souri,et al.  Resource Management Approaches in Fog Computing: a Comprehensive Review , 2019, Journal of Grid Computing.