Energy Minimization in D2D-Assisted Cache-Enabled Internet of Things: A Deep Reinforcement Learning Approach

Mobile edge caching (MEC) and device-to-device (D2D) communications are two potential technologies to resolve traffic overload problems in the Internet of Things. Previous works usually investigate them separately with MEC for traffic offloading and D2D for information transmission. In this article, a joint framework consisting of MEC and cache-enabled D2D communications is proposed to minimize the energy cost of systematic traffic transmission, where file popularity and user preference are the critical criteria for small base stations (SBSs) and user devices, respectively. Under this framework, we propose a novel caching strategy, where the Markov decision process is applied to model the requesting behaviors. A novel scheme based on reinforcement learning (RL) is proposed to reveal the popularity of files as well as users’ preference. In particular, a $Q$-learning algorithm and a deep $Q$-network algorithm are, respectively, applied to user devices and the SBS due to different complexities of status. To save the energy cost of systematic traffic transmission, users acquire partial traffic through D2D communications based on the cached contents and user distribution. Taking the memory limits, D2D available files, and status changing into consideration, the proposed RL algorithm enables user devices and the SBS to prefetch the optimal files while learning, which can reduce the energy cost significantly. Simulation results demonstrate the superior energy saving performance of the proposed RL-based algorithm over other existing methods under various conditions.

[1]  Tao Chen,et al.  Device-To-Device (D2D) Communication in Cellular Network - Performance Analysis of Optimum and Practical Communication Mode Selection , 2010, 2010 IEEE Wireless Communication and Networking Conference.

[2]  Zibin Zheng,et al.  Online Deep Reinforcement Learning for Computation Offloading in Blockchain-Empowered Mobile Edge Computing , 2019, IEEE Transactions on Vehicular Technology.

[3]  Victor C. M. Leung,et al.  Joint computation and communication resource allocation in mobile-edge cloud computing networks , 2016, 2016 IEEE International Conference on Network Infrastructure and Digital Content (IC-NIDC).

[4]  Alireza Sadeghi,et al.  Optimal and Scalable Caching for 5G Using Reinforcement Learning of Space-Time Popularities , 2017, IEEE Journal of Selected Topics in Signal Processing.

[5]  Ren Ping Liu,et al.  ResInNet: A Novel Deep Neural Network With Feature Reuse for Internet of Things , 2019, IEEE Internet of Things Journal.

[6]  Alex Graves,et al.  Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.

[7]  Dong Liu,et al.  Energy Efficiency of Downlink Networks With Caching at Base Stations , 2015, IEEE Journal on Selected Areas in Communications.

[8]  Senem Velipasalar,et al.  Learning-Based Delay-Aware Caching in Wireless D2D Caching Networks , 2018, IEEE Access.

[9]  Xiaofei Wang,et al.  Cache in the air: exploiting content caching and delivery techniques for 5G systems , 2014, IEEE Communications Magazine.

[10]  Alexandros G. Dimakis,et al.  Femtocaching and device-to-device collaboration: A new architecture for wireless video distribution , 2012, IEEE Communications Magazine.

[11]  Xiaoxiang Wang,et al.  On the Design of Computation Offloading in Cache-Aided D2D Multicast Networks , 2018, IEEE Access.

[12]  Ying Wang,et al.  Clustered device-to-device caching based on file preferences , 2016, 2016 IEEE 27th Annual International Symposium on Personal, Indoor, and Mobile Radio Communications (PIMRC).

[13]  Min Sheng,et al.  Learning-Based Content Caching and Sharing for Wireless Networks , 2017, IEEE Transactions on Communications.

[14]  Zixiang Xiong,et al.  Optimal Caching and Scheduling for Cache-Enabled D2D Communications , 2017, IEEE Communications Letters.

[15]  Guan Gui,et al.  Deep Learning for an Effective Nonorthogonal Multiple Access Scheme , 2018, IEEE Transactions on Vehicular Technology.

[16]  Justin P. Coon,et al.  Optimal Routing for Multihop Social-Based D2D Communications in the Internet of Things , 2018, IEEE Internet of Things Journal.

[17]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[18]  Suhas N. Diggavi,et al.  Content caching and delivery over heterogeneous wireless networks , 2014, 2015 IEEE Conference on Computer Communications (INFOCOM).

[19]  Mihaela van der Schaar,et al.  Popularity-driven content caching , 2016, IEEE INFOCOM 2016 - The 35th Annual IEEE International Conference on Computer Communications.

[20]  Alec Radford,et al.  Proximal Policy Optimization Algorithms , 2017, ArXiv.

[21]  Naixue Xiong,et al.  UAV Autonomous Target Search Based on Deep Reinforcement Learning in Complex Disaster Scene , 2019, IEEE Access.

[22]  Hyoung-Nam Kim,et al.  Deep Learning Detection in MIMO Decode-Forward Relay Channels , 2019, IEEE Access.

[23]  Justin P. Coon,et al.  Performance Analysis for Multihop Full-Duplex IoT Networks Subject to Poisson Distributed Interferers , 2019, IEEE Internet of Things Journal.

[24]  Victor C. M. Leung,et al.  Integrated Computing, Caching, and Communication for Trust-Based Social Networks: A Big Data DRL Approach , 2018, 2018 IEEE Global Communications Conference (GLOBECOM).

[25]  Bhaskar Krishnamachari,et al.  Deep Reinforcement Learning for Dynamic Multichannel Access in Wireless Networks , 2018, IEEE Transactions on Cognitive Communications and Networking.

[26]  Fumiyuki Adachi,et al.  Transceiver Design and Multihop D2D for UAV IoT Coverage in Disasters , 2019, IEEE Internet of Things Journal.

[27]  Alexandros Nanopoulos,et al.  Modeling Users Preference Dynamics and Side Information in Recommender Systems , 2016, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[28]  Mehdi Bennis,et al.  A transfer learning approach for cache-enabled wireless networks , 2015, 2015 13th International Symposium on Modeling and Optimization in Mobile, Ad Hoc, and Wireless Networks (WiOpt).

[29]  Yang Yi,et al.  Energy Harvesting-Based D2D-Assisted Machine-Type Communications , 2017, IEEE Transactions on Communications.

[30]  Theodore S. Rappaport,et al.  Wireless communications - principles and practice , 1996 .

[31]  Haitao Wang,et al.  Deep reinforcement learning with experience replay based on SARSA , 2016, 2016 IEEE Symposium Series on Computational Intelligence (SSCI).

[32]  Jifeng Guo,et al.  A Deep Q-network (DQN) Based Path Planning Method for Mobile Robots , 2018, 2018 IEEE International Conference on Information and Automation (ICIA).

[33]  Shuguang Cui,et al.  Joint offloading and computing optimization in wireless powered mobile-edge computing systems , 2017, 2017 IEEE International Conference on Communications (ICC).

[34]  Qi Shi,et al.  A Deep Learning Approach to Network Intrusion Detection , 2018, IEEE Transactions on Emerging Topics in Computational Intelligence.

[35]  Chenyang Yang,et al.  Caching Policy for Cache-Enabled D2D Communications by Learning User Preference , 2017, IEEE Transactions on Communications.

[36]  Rudolf Mathar,et al.  Deep Reinforcement Learning based Resource Allocation in Low Latency Edge Computing Networks , 2018, 2018 15th International Symposium on Wireless Communication Systems (ISWCS).

[37]  Robert W. Heath,et al.  Five disruptive technology directions for 5G , 2013, IEEE Communications Magazine.

[38]  Yim-Fun Hu,et al.  Saving Energy in Mobile Devices Using Mobile Device Cloudlet in Mobile Edge Computing for 5G , 2017, 2017 IEEE International Conference on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber, Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData).

[39]  Chenyang Yang,et al.  Caching in Base Station with Recommendation via Q-Learning , 2017, 2017 IEEE Wireless Communications and Networking Conference (WCNC).

[40]  M. Draief,et al.  Placing dynamic content in caches with small population , 2016, IEEE INFOCOM 2016 - The 35th Annual IEEE International Conference on Computer Communications.

[41]  Fang Dong,et al.  Power Consumption Minimization in Cache-Enabled Mobile Networks , 2019, IEEE Transactions on Vehicular Technology.