Deep Q-Learning for Self-Organizing Networks Fault Management and Radio Performance Improvement

We propose an algorithm to automate fault management in an outdoor cellular network using deep reinforcement learning (RL) against wireless impairments. This algorithm enables the cellular network cluster to self-heal by allowing RL to learn how to improve the downlink signal to interference plus noise ratio through exploration and exploitation of various alarm corrective actions. The main contributions of this paper are to 1) introduce a deep RL-based fault handling algorithm which self-organizing networks can implement in a polynomial runtime and 2) show that this fault management method can improve the radio link performance in a realistic network setup. Simulation results show that our proposed algorithm learns an action sequence to clear alarms and improve the performance in the cellular cluster better than existing algorithms, even against the randomness of the network fault occurrences and user movements.

[1]  Richard S. Sutton,et al.  Introduction to Reinforcement Learning , 1998 .

[2]  Marc'Aurelio Ranzato,et al.  Large Scale Distributed Deep Networks , 2012, NIPS.

[3]  Raquel Barco,et al.  Data mining for fuzzy diagnosis systems in LTE networks , 2015, Expert Syst. Appl..

[4]  Victor C. M. Leung,et al.  Deep-Reinforcement-Learning-Based Optimization for Cache-Enabled Opportunistic Interference Alignment Wireless Networks , 2017, IEEE Transactions on Vehicular Technology.

[5]  Markus Rupp,et al.  The Vienna LTE-Advanced Simulators: Up and Downlink, Link and System Level Simulation , 2016 .

[6]  Andreas Mitschele-Thiel,et al.  Cognitive Cellular Networks: A Q-Learning Framework for Self-Organizing Networks , 2016, IEEE Transactions on Network and Service Management.

[7]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[8]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[9]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[10]  Yoshikazu Miyanaga,et al.  Multi-agent Q-learning for autonomous D2D communication , 2016, 2016 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS).

[11]  Muhammad Ali Imran,et al.  Challenges in 5G: how to empower SON with big data for enabling 5G , 2014, IEEE Network.

[12]  Masashi Sugiyama,et al.  Deep Reinforcement Learning with Relative Entropy Stochastic Search , 2017 .

[13]  Markus Rupp,et al.  The Vienna LTE-Advanced Simulators , 2016 .

[14]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[15]  Selahattin Kosunalp A New Energy Prediction Algorithm for Energy-Harvesting Wireless Sensor Networks With Q-Learning , 2016, IEEE Access.

[16]  Theodore S. Rappaport,et al.  Millimeter-wave distance-dependent large-scale propagation measurements and path loss models for outdoor and indoor 5G systems , 2015, 2016 10th European Conference on Antennas and Propagation (EuCAP).

[17]  David Tse,et al.  Fundamentals of Wireless Communication , 2005 .

[18]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.