Self-Tuning Sectorization: Deep Reinforcement Learning Meets Broadcast Beam Optimization

Beamforming in multiple input multiple output (MIMO) systems is one of the key technologies for modern wireless communication. Creating appropriate sector-specific broadcast beams are essential for enhancing the coverage of cellular network and for improving the broadcast operation for control signals. However, in order to maximize the coverage, patterns for broadcast beams need to be adapted based on the users’ distribution and movement over time. In this work, we present self-tuning sectorization: a deep reinforcement learning framework to optimize MIMO broadcast beams autonomously and dynamically based on users’ distribution in the network. Taking directly UE measurement results as input, deep reinforcement learning agent can track and predict the UE distribution pattern and come up with the best broadcast beams for each cell. Extensive simulation results show that the introduced framework can achieve the optimal coverage, and converge to the oracle solution for both single sector and multiple sectors environment, and for both periodic and Markov mobility patterns.

[1]  Zhu Han,et al.  Self-Organization in Small Cell Networks: A Reinforcement Learning Approach , 2013, IEEE Transactions on Wireless Communications.

[2]  Chung-Ju Chang,et al.  Q-learning-based multirate transmission control scheme for RRM in multimedia WCDMA systems , 2004, IEEE Transactions on Vehicular Technology.

[3]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[4]  David Silver,et al.  Deep Reinforcement Learning with Double Q-Learning , 2015, AAAI.

[5]  Agostino Poggi,et al.  Multiagent Systems , 2006, Intelligenza Artificiale.

[6]  Yik-Chung Wu,et al.  DoA Estimation and Capacity Analysis for 3-D Millimeter Wave Massive-MIMO/FD-MIMO OFDM Systems , 2016, IEEE Transactions on Wireless Communications.

[7]  Ana Galindo-Serrano,et al.  Distributed Q-Learning for Aggregated Interference Control in Cognitive Radio Networks , 2010, IEEE Transactions on Vehicular Technology.

[8]  Robert W. Heath,et al.  Five disruptive technology directions for 5G , 2013, IEEE Communications Magazine.

[9]  Lingjia Liu,et al.  Multi-Cell Multi-User Massive FD-MIMO: Downlink Precoding and Throughput Analysis , 2018, IEEE Transactions on Wireless Communications.

[10]  Jeffrey G. Andrews,et al.  What Will 5G Be? , 2014, IEEE Journal on Selected Areas in Communications.

[11]  Thomas L. Marzetta,et al.  Noncooperative Cellular Wireless with Unlimited Numbers of Base Station Antennas , 2010, IEEE Transactions on Wireless Communications.

[12]  S. Haykin,et al.  A Q-learning-based dynamic channel assignment technique for mobile communication systems , 1999 .

[13]  A. Lozano,et al.  What Will 5 G Be ? , 2014 .

[14]  Yan Li,et al.  Angle and Delay Estimation for 3-D Massive MIMO/FD-MIMO Systems Based on Parametric Channel Modeling , 2017, IEEE Transactions on Wireless Communications.

[15]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[16]  Boon Loong Ng,et al.  Full dimension mimo (FD-MIMO): the next evolution of MIMO in LTE systems , 2014, IEEE Wireless Communications.

[17]  Haibo He,et al.  Distributive Dynamic Spectrum Access Through Deep Reinforcement Learning: A Reservoir Computing-Based Approach , 2018, IEEE Internet of Things Journal.

[18]  Xiqi Gao,et al.  Cellular architecture and key technologies for 5G wireless communication networks , 2014, IEEE Communications Magazine.

[19]  Krishna Sayana,et al.  Downlink MIMO in LTE-advanced: SU-MIMO vs. MU-MIMO , 2012, IEEE Communications Magazine.

[20]  Yang Yang,et al.  Self-configuration and self-optimization for LTE networks , 2010, IEEE Communications Magazine.

[21]  Gerhard Weiss,et al.  Multiagent Systems , 1999 .

[22]  Boon Loong Ng,et al.  Full-dimension MIMO (FD-MIMO) for next generation cellular technology , 2013, IEEE Communications Magazine.

[23]  Long-Ji Lin,et al.  Reinforcement learning for robots using neural networks , 1992 .

[24]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[25]  Rakesh Misra,et al.  Towards Self-Driving Radios: Physical-Layer Control using Deep Reinforcement Learning , 2019, HotMobile.

[26]  Dong Liang,et al.  Self-configuration and self-optimization in LTE-advanced heterogeneous networks , 2013, IEEE Communications Magazine.

[27]  Henning Sanneck,et al.  LTE Self-Organising Networks (SON): Network Management Automation for Operational Efficiency , 2012 .

[28]  Barbara Messing,et al.  An Introduction to MultiAgent Systems , 2002, Künstliche Intell..

[29]  Tamer A. ElBatt,et al.  Distributed Cooperative Q-Learning for Power Allocation in Cognitive Femtocell Networks , 2012, 2012 IEEE Vehicular Technology Conference (VTC Fall).

[30]  Juan Sanchez-Gonzalez,et al.  A roadmap from UMTS optimization to LTE self-optimization , 2011, IEEE Communications Magazine.