QoS-Driven Scheduling in 5G Radio Access Networks - A Reinforcement Learning Approach

The expected diversity of services and the variety of use cases in 5G networks will require a flexible Radio Resource Management able to satisfy the heterogeneous Quality of Service (QoS) requirements. Classical scheduling strategies have been designed to deal mainly with some particular QoS requirements for specific traffic types. To improve the scheduling performance, this paper proposes an innovative scheduler framework, that selects at each transmission time interval, the appropriate scheduling strategy capable to maximize the users$'$ satisfaction measure in terms of distinct QoS requirements. Neural networks and the Reinforcement Learning paradigm are jointly used to learn the best scheduling decision based on the past experiences. The simulation results show very good convergence properties for the proposed policies, and notable QoS improvements with the respect to the baseline scheduling solutions.

[1]  Sudharman K. Jayaweera,et al.  A Survey on Machine-Learning Techniques in Cognitive Radios , 2013, IEEE Communications Surveys & Tutorials.

[2]  Ovidiu Iacoboaiea,et al.  SON Coordination in Heterogeneous Networks: A Reinforcement Learning Framework , 2016, IEEE Transactions on Wireless Communications.

[3]  Giuseppe Piro,et al.  Downlink Packet Scheduling in LTE Cellular Networks: Key Design Issues and a Survey , 2013, IEEE Communications Surveys & Tutorials.

[4]  A. Lozano,et al.  What Will 5 G Be ? , 2014 .

[5]  Valentin Savin,et al.  Backhaul-aware small cell DTX based on fuzzy Q-Learning in heterogeneous cellular networks , 2016, 2016 IEEE International Conference on Communications (ICC).

[6]  Giuseppe Piro,et al.  Simulating LTE Cellular Systems: An Open-Source Framework , 2011, IEEE Transactions on Vehicular Technology.

[7]  Chung-Ju Chang,et al.  An Intelligent Priority Resource Allocation Scheme for LTE-A Downlink Systems , 2012, IEEE Wireless Communications Letters.

[8]  Euhanna Ghadimi,et al.  A reinforcement learning approach to power control and rate adaptation in cellular networks , 2016, 2017 IEEE International Conference on Communications (ICC).

[9]  Robert Babuska,et al.  A Survey of Actor-Critic Reinforcement Learning: Standard and Natural Policy Gradients , 2012, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[10]  Jianping Chen,et al.  Adaptive proportional fair parameterization based LTE scheduling using continuous actor-critic reinforcement learning , 2014, 2014 IEEE Global Communications Conference.

[11]  A. Benjebbour,et al.  Design considerations for a 5G network architecture , 2014, IEEE Communications Magazine.