Deep Reinforcement Learning for Dynamic Access Control with Battery Prediction for Mobile-Edge Computing in Green IoT Networks

Mobile Edge Computing (MEC) technology has emerged as a promising paradigm to reduce the energy consumption for the resource-constrained and energy-limited Internet of Things (IoT) networks. In this paper, benefiting from energy harvesting technique (EH), we study the dynamic MEC-access control problem for maximizing the long-term average uplink transmission rate whilst minimizing the transmission energy consumption for green IoT networks, in which the IoT device is powered by a rechargeable battery that can harvest energy from the surrounding environments. In particular, this problem is formulated as a Markov decision process with system dynamics unknown. On accounting of the dynamics of the wireless channel state, the energy arrival, and the mobility of the IoT device, a Long Short-Term Memory (LSTM) enhanced Deep Q-Network (DQN) based (LSDQN) access control algorithm is proposed for the IoT network. In the proposed algorithm, the LSTM model is used to predict the battery status for assisting the IoT device to determine the optimal access control decision by DQN with the target of maximizing the average uplink rate whilst minimizing the energy consumption. Finally, extensive simulation results verify the performance of the proposed algorithm.

[1]  Kaibin Huang,et al.  Exploiting Non-Causal CPU-State Information for Energy-Efficient Mobile Cooperative Computing , 2017, IEEE Transactions on Wireless Communications.

[2]  Xuemin Shen,et al.  Self-Organized Energy Management in Energy Harvesting Small Cell Networks , 2018, 2018 IEEE Global Communications Conference (GLOBECOM).

[3]  Xiong Xiong,et al.  Joint Computation Offloading and Multiuser Scheduling Using Approximate Dynamic Programming in NB-IoT Edge Computing System , 2019, IEEE Internet of Things Journal.

[4]  Jie Zhang,et al.  Energy-Aware Computation Offloading and Transmit Power Allocation in Ultradense IoT Networks , 2019, IEEE Internet of Things Journal.

[5]  Tarik Taleb,et al.  Survey on Multi-Access Edge Computing for Internet of Things Realization , 2018, IEEE Communications Surveys & Tutorials.

[6]  Shuguang Cui,et al.  Reinforcement Learning-Based Multiaccess Control and Battery Prediction With Energy Harvesting in IoT Systems , 2018, IEEE Internet of Things Journal.

[7]  Laizhong Cui,et al.  Joint Optimization of Energy Consumption and Latency in Mobile Edge Computing for Internet of Things , 2019, IEEE Internet of Things Journal.

[8]  Duy Trong Ngo,et al.  A Distributed Energy-Harvesting-Aware Routing Algorithm for Heterogeneous IoT Networks , 2018, IEEE Transactions on Green Communications and Networking.

[9]  Tobias Weber,et al.  Reinforcement Learning for Energy Harvesting Decode-and-Forward Two-Hop Communications , 2017, IEEE Transactions on Green Communications and Networking.

[10]  Yan Zhang,et al.  Joint Computation Offloading and User Association in Multi-Task Mobile Edge Computing , 2018, IEEE Transactions on Vehicular Technology.