Double Deep Q-Network for Power Allocation in Cloud Radio Access Network

Cloud radio access network (CRAN) facilitates resource allocation (RA) by isolating remote radio heads (RRHs) from baseband units (BBUs). Traditional RA algorithms save energy by dynamically turning on/off RRHs and allocating power in each time slot. However, when the energy switching cost is considered, the decisions of turning on/off RRHs in adjacent time slots are correlated, which cannot be solved directly. Fortunately, deep reinforcement learning (DRL) can effectively model such problem, which motivates us to minimize the total power consumption subject to the constraints on per-RRH transmit power and user rates. Our starting point is the deep Q network (DQN), which is a combination of a neural network and Q-learning. In each time slot, DQN turns on /off a RRH yielding the largest Q-value (known as action value) prior to solving a power minimization problem for active RRHs. However, DQN yields Q-value overestimation issue, which stems from using the same network to choose the best action and to compute the target Qvalue of taking that action at the next state. To further increase the CRAN power savings, we propose a Double DQN-based framework by decoupling the action selection from the target Q-value generation. Simulation results indicate that the Double DQN-based RA method outperforms the DQN-based RA algorithm in terms of total power consumption.

[1]  Geoffrey Ye Li,et al.  Deep Reinforcement Learning Based Resource Allocation for V2V Communications , 2018, IEEE Transactions on Vehicular Technology.

[2]  Lisa Turner,et al.  Applications of Second Order Cone Programming , 2012 .

[3]  Tiejun Lv,et al.  Deep Q-Learning Based Dynamic Resource Allocation for Self-Powered Ultra-Dense Networks , 2018, 2018 IEEE International Conference on Communications Workshops (ICC Workshops).

[4]  Min Sheng,et al.  Utility-Based Resource Allocation for Multi-Channel Decentralized Networks , 2014, IEEE Transactions on Communications.

[5]  Jing Wang,et al.  A deep reinforcement learning based framework for power-efficient resource allocation in cloud RANs , 2017, 2017 IEEE International Conference on Communications (ICC).

[6]  Chien-Hwa Hwang,et al.  ON 5G RADIO ACCESS ARCHITECTURE AND TECHNOLOGY , 2015 .

[7]  Rui Zhang,et al.  Downlink and Uplink Energy Minimization Through User Association and Beamforming in C-RAN , 2014, IEEE Transactions on Wireless Communications.

[8]  Gerhard Fettweis,et al.  Are Heterogeneous Cloud-Based Radio Access Networks Cost Effective? , 2015, IEEE Journal on Selected Areas in Communications.

[9]  David Silver,et al.  Deep Reinforcement Learning with Double Q-Learning , 2015, AAAI.

[10]  Vijay K. Bhargava,et al.  Green Cellular Networks: A Survey, Some Research Issues and Challenges , 2011, IEEE Communications Surveys & Tutorials.

[11]  Ami Wiesel,et al.  Linear precoding via conic optimization for fixed MIMO receivers , 2006, IEEE Transactions on Signal Processing.

[12]  Yoong Choon Chang,et al.  Deep Reinforcement Learning for Resource Allocation in 5G Communications , 2019, 2019 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC).

[13]  Sanjeev Jain,et al.  Green Communication in Next Generation Cellular Networks: A Survey , 2017, IEEE Access.

[14]  Minoru Etoh,et al.  Energy Consumption Issues on Mobile Network Systems , 2008, 2008 International Symposium on Applications and the Internet.

[15]  P ? ? ? ? ? ? ? % ? ? ? ? , 1991 .

[16]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[17]  Su Fong Chien,et al.  Energy-Efficient Power Allocation for Distributed Antenna Systems With Proportional Fairness , 2017, IEEE Transactions on Green Communications and Networking.

[18]  Eduardo F. Morales,et al.  An Introduction to Reinforcement Learning , 2011 .

[19]  Wei Chen,et al.  Enhanced Group Sparse Beamforming for Green Cloud-RAN: A Random Matrix Approach , 2017, IEEE Transactions on Wireless Communications.

[20]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[21]  Yuanming Shi,et al.  Group Sparse Beamforming for Green Cloud-RAN , 2013, IEEE Transactions on Wireless Communications.

[22]  Yuan Li,et al.  Heterogeneous cloud radio access networks: a new perspective for enhancing spectral and energy efficiencies , 2014, IEEE Wireless Communications.

[23]  Tiejun Lv,et al.  Deep reinforcement learning based computation offloading and resource allocation for MEC , 2018, 2018 IEEE Wireless Communications and Networking Conference (WCNC).

[24]  Wei Yu,et al.  Energy Efficiency of Downlink Transmission Strategies for Cloud Radio Access Networks , 2016, IEEE Journal on Selected Areas in Communications.

[25]  Hamed Haddadi,et al.  Deep Learning in Mobile and Wireless Networking: A Survey , 2018, IEEE Communications Surveys & Tutorials.