Resource Allocation in Information-Centric Wireless Networking With D2D-Enabled MEC: A Deep Reinforcement Learning Approach

Recently, information-centric wireless networks (ICWNs) have become a promising Internet architecture of the next generation, which allows network nodes to have computing and caching capabilities and adapt to the growing mobile data traffic in 5G high-speed communication networks. However, the design of ICWN is still faced with various challenges with respect to capacity and traffic. Therefore, mobile edge computing (MEC) and device-to-device (D2D) communications can be employed to aid offloading the core networks. This paper investigates the optimal policy for resource allocation in ICWNs by maximizing the spectrum efficiency and system capacity of the overall network. Due to unknown and stochastic properties of the wireless channel environment, this problem was modeled as a Markov decision process. In continuous-valued state and action variables, the policy gradient approach was employed to learn the optimal policy through interactions with the environment. We first recognized the communication mode according to the location of the cached content, considering whether it is D2D mode or cellular mode. Then, we adopt the Gaussian distribution as the parameterization strategy to generate continuous stochastic actions to select power. In addition, we use softmax to output channel selection to maximize system capacity and spectrum efficiency while avoiding interference to cellular users. The numerical experiments show that our learning method performs well in a D2D-enabled MEC system.

[1]  Min Zhao,et al.  Power Control for D2D Communication Using Multi-Agent Reinforcement Learning , 2018, 2018 IEEE/CIC International Conference on Communications in China (ICCC).

[2]  Wenbo Wang,et al.  Interference avoidance mechanisms in the hybrid cellular and device-to-device systems , 2009, 2009 IEEE 20th International Symposium on Personal, Indoor and Mobile Radio Communications.

[3]  Xianzhong Xie,et al.  Intelligent Resource Management Based on Reinforcement Learning for Ultra-Reliable and Low-Latency IoV Communication Networks , 2019, IEEE Transactions on Vehicular Technology.

[4]  Victor C. M. Leung,et al.  Integrated Computing, Caching, and Communication for Trust-Based Social Networks: A Big Data DRL Approach , 2018, 2018 IEEE Global Communications Conference (GLOBECOM).

[5]  Victor C. M. Leung,et al.  An Information Centric Networking approach towards contextualized edge service , 2015, 2015 12th Annual IEEE Consumer Communications and Networking Conference (CCNC).

[6]  Liang Liu,et al.  Joint Task Assignment and Resource Allocation for D2D-Enabled Mobile-Edge Computing , 2019, IEEE Transactions on Communications.

[7]  Khaled Ben Letaief,et al.  Mobility-aware caching for content-centric wireless networks: modeling and methodology , 2016, IEEE Communications Magazine.

[8]  Gang Feng,et al.  Proactive content caching by exploiting transfer learning for mobile edge computing , 2018, Int. J. Commun. Syst..

[9]  Andrey Koucheryavy,et al.  5G framework based on multi-level edge computing with D2D enabled communication , 2018, 2018 20th International Conference on Advanced Communication Technology (ICACT).

[10]  Xiaojiang Du,et al.  Self-healing sensor networks with distributed decision making , 2007, Int. J. Sens. Networks.

[11]  Xiang Cheng,et al.  Distributed resource allocation for device-to-device communications underlaying cellular networks , 2013, 2013 IEEE International Conference on Communications (ICC).

[12]  Kin K. Leung,et al.  Online Placement of Multi-Component Applications in Edge Computing Environments , 2016, IEEE Access.

[13]  Alex Graves,et al.  Recurrent Models of Visual Attention , 2014, NIPS.

[14]  F. Richard Yu,et al.  Resource Allocation for Information-Centric Virtualized Heterogeneous Networks With In-Network Caching and Mobile Edge Computing , 2017, IEEE Transactions on Vehicular Technology.

[15]  Qianbin Chen,et al.  Computation Offloading and Resource Allocation in Wireless Cellular Networks With Mobile Edge Computing , 2017, IEEE Transactions on Wireless Communications.

[16]  Mohsen Guizani,et al.  Monopolistic Models for Resource Allocation: A Probabilistic Reinforcement Learning Approach , 2018, IEEE Access.

[17]  Rose Qingyang Hu,et al.  Mobility-Aware Edge Caching and Computing in Vehicle Networks: A Deep Reinforcement Learning , 2018, IEEE Transactions on Vehicular Technology.

[18]  Xi Zhang,et al.  Information-centric network function virtualization over 5g mobile wireless networks , 2015, IEEE Network.

[19]  Zhu Han,et al.  Game-theoretic resource allocation methods for device-to-device communication , 2014, IEEE Wireless Communications.

[20]  Rong Chai,et al.  Task Execution Cost Minimization-Based Joint Computation Offloading and Resource Allocation for Cellular D2D MEC Systems , 2019, IEEE Systems Journal.

[21]  Chau Yuen,et al.  Fast Video Frame Correlation Analysis for Vehicular Networks by Using CVS–CNN , 2019, IEEE Transactions on Vehicular Technology.

[22]  Gang Xie,et al.  Joint Mode Selection and Power Adaptation for D2D Communication with Reinforcement Learning , 2018, 2018 15th International Symposium on Wireless Communication Systems (ISWCS).

[23]  Suyong Eum,et al.  Information-Centric Networking (ICN) Research Challenges , 2016, RFC.

[24]  Zhao Ming,et al.  Q-learning based power control algorithm for D2D communication , 2016 .

[25]  Mohsen Guizani,et al.  A Routing-Driven Key Management Scheme for Heterogeneous Sensor Networks , 2007, 2007 IEEE International Conference on Communications.