Joint resource allocation and power control for D2D communication with deep reinforcement learning in MCC

Abstract Mission-critical communication (MCC) is one of the main goals in 5G, which can leverage multiple device-to-device (D2D) connections to enhance reliability for mission-critical communication. In MCC, D2D users can reuses the non-orthogonal wireless resources of cellular users without a base station (BS). Meanwhile, the D2D users will generate co-channel interference to cellular users and hence affect their quality-of-service (QoS). To comprehensively improve the user experience, we proposed a novel approach, which embraces resource allocation and power control along with Deep Reinforcement Learning (DRL). In this paper, multiple procedures are carefully designed to assist in developing our proposal. As a starter, a scenario with multiple D2D pairs and cellular users in a cell will be modeled; followed by the analysis of issues pertaining to resource allocation and power control as well as the formulation of our optimization goal; and finally, a DRL method based on spectrum allocation strategy will be created, which can ensure D2D users to obtain the sufficient resource for their QoS improvement. With the resource data provided, which D2D users capture by interacting with surroundings, the DRL method can help the D2D users autonomously selecting an available channel and power to maximize system capacity and spectrum efficiency while minimizing interference to cellular users. Experimental results show that our learning method performs well to improve resource allocation and power control significantly.

[1]  Kobi Cohen,et al.  Deep Multi-User Reinforcement Learning for Dynamic Spectrum Access in Multichannel Wireless Networks , 2017, GLOBECOM 2017 - 2017 IEEE Global Communications Conference.

[2]  Xiaojiang Du,et al.  Self-healing sensor networks with distributed decision making , 2007, Int. J. Sens. Networks.

[3]  Tarcisio F. Maciel,et al.  Radio Resource Management for Device-to-Device Communications in Long Term Evolution Networks , 2014 .

[4]  Dan Wang,et al.  From IoT to 5G I-IoT: The Next Generation IoT-Based Intelligent Algorithms and 5G Technologies , 2018, IEEE Communications Magazine.

[5]  Sang Quang Nguyen,et al.  Social-Aware Caching and Resource Sharing Maximized Video Delivery Capacity in 5G Ultra-Dense Networks , 2019, Mob. Networks Appl..

[6]  Mohsen Guizani,et al.  5G D2D Networks: Techniques, Challenges, and Future Prospects , 2018, IEEE Systems Journal.

[7]  Yan Chen,et al.  Intelligent 5G: When Cellular Networks Meet Artificial Intelligence , 2017, IEEE Wireless Communications.

[8]  Zhu Han,et al.  Game-theoretic resource allocation methods for device-to-device communication , 2014, IEEE Wireless Communications.

[9]  Weidong Wang,et al.  Deep Reinforcement Learning Based Dynamic Channel Allocation Algorithm in Multibeam Satellite Systems , 2018, IEEE Access.

[10]  Zhu Han,et al.  Coalitional Games for Resource Allocation in the Device-to-Device Uplink Underlaying Cellular Networks , 2014, IEEE Transactions on Wireless Communications.

[11]  Ying-Chang Liang,et al.  Applications of Deep Reinforcement Learning in Communications and Networking: A Survey , 2018, IEEE Communications Surveys & Tutorials.

[12]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[13]  Mohsen Guizani,et al.  Monopolistic Models for Resource Allocation: A Probabilistic Reinforcement Learning Approach , 2018, IEEE Access.