Applying Event-Triggered Q-Learning Algorithms

2 ||| 1556-6072/19©2019IEEE IEEE VEHICULAR TECHNOLOGY MAGAZINE | MONTH 2020 Ultradense networks (UDNs) have emerged as a promising architecture that can support the extremely high demand for data traffic in the future. Through the dense deployment of massive small base stations (SBSs), the system can be well promoted in terms of network capacity and spectrum efficiency, but this deployment scheme will also bring huge challenges to wireless resource management. Artificial intelligence (AI) can be applied to UDN scenarios, enabling intelligent communication devices to learn and complete resource allocation. This article discusses resource allocation schemes based on AI algorithms in UDNs and proposes an event-triggered, reinforcement-learning-based subchannel and power allocation algorithm. We consider nonorthogonal multiple access (NOMA) in UDNs, which allows a subchannel to be used by multiple users at the same time. Each user is regarded as an agent, and each agent obtains observational information from the environment during the learning process. We design event-triggered conditions based on current and previous moments of observational information, and the agent uses those conditions to decide whether to update policies and perform actions. Simulation results show the effectiveness of the proposed eventtriggered, reinforcement-learning-based resource allocation algorithm.

[1]  Victor C. M. Leung,et al.  Energy Efficient User Association and Power Allocation in Millimeter-Wave-Based Ultra Dense Networks With Energy Harvesting Base Stations , 2017, IEEE Journal on Selected Areas in Communications.

[2]  Zhimin Zeng,et al.  Exploiting Polarization for System Capacity Maximization in Ultra-Dense Small Cell Networks , 2017, IEEE Access.

[3]  Chunlong He,et al.  Power Allocation Schemes Based on Machine Learning for Distributed Antenna Systems , 2019, IEEE Access.

[4]  Kentaro Ishizu,et al.  Big Data Analytics, Machine Learning, and Artificial Intelligence in Next-Generation Wireless Networks , 2017, IEEE Access.

[5]  Nikos D. Sidiropoulos,et al.  Learning to optimize: Training deep neural networks for wireless resource management , 2017, 2017 IEEE 18th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC).

[6]  Wen Wang,et al.  A Cluster-Based Energy-Efficient Resource Management Scheme for Ultra-Dense Networks , 2016, IEEE Access.

[7]  Ming Xiao,et al.  Heterogeneous Ultradense Networks with NOMA: System Architecture, Coordination Framework, and Performance Evaluation , 2018, IEEE Vehicular Technology Magazine.

[8]  Zhu Han,et al.  Machine Learning Paradigms for Next-Generation Wireless Networks , 2017, IEEE Wireless Communications.

[9]  Amr M. Youssef,et al.  Ultra-Dense Networks: A Survey , 2016, IEEE Communications Surveys & Tutorials.

[10]  Shin-Ming Cheng,et al.  eNB Selection for Machine Type Communications Using Reinforcement Learning Based Markov Decision Process , 2017, IEEE Transactions on Vehicular Technology.

[11]  George K. Karagiannidis,et al.  Resource Allocation in NOMA-Based Fog Radio Access Networks , 2018, IEEE Wireless Communications.

[12]  Pablo Soldati,et al.  Spectrum and Network Density Management in 5G Ultra-Dense Networks , 2017, IEEE Wireless Communications.

[13]  Zhijin Qin,et al.  User Association and Resource Allocation in Unified NOMA Enabled Heterogeneous Ultra Dense Networks , 2018, IEEE Communications Magazine.

[14]  Geoffrey Ye Li,et al.  Power of Deep Learning for Channel Estimation and Signal Detection in OFDM Systems , 2017, IEEE Wireless Communications Letters.

[15]  Tao Jiang,et al.  Ultra-Dense Heterogeneous Relay Networks: A Non-Uniform Traffic Hotspot Case , 2017, IEEE Network.