Mobile Edge Computing Against Smart Attacks with Deep Reinforcement Learning in Cognitive MIMO IoT Systems

In wireless Internet of Things (IoT) systems, the multi-input multi-output (MIMO) and cognitive radio (CR) techniques are usually involved into the mobile edge computing (MEC) structure to improve the spectrum efficiency and transmission reliability. However, such a CR based MIMO IoT system will suffer from a variety of smart attacks from wireless environments, even the MEC servers in IoT systems are not secure enough and vulnerable to these attacks. In this paper, we investigate a secure communication problem in a cognitive MIMO IoT system comprising of a primary user (PU), a secondary user (SU), a smart attacker and several MEC servers. The target of our system design is to optimize utility of the SU, including its efficiency and security. The SU will choose an idle MEC server that is not occupied by the PU in the CR scenario, and allocates a proper offloading rate of its computation tasks to the server, by unloading such tasks with proper transmit power. In such a CR IoT system, the attacker will select one type of smart attacks. Then two deep reinforcement learning based resource allocation strategies are proposed to find an optimal policy of maximal utility without channel state information(CSI), one of which is the Dyna architecture and Prioritized sweeping based Edge Server Selection (DPESS) strategy, and the other is the Deep Q-network based Edge Server Selection (DESS) strategy. Specifically, the convergence speed of the DESS scheme is significantly improved due to the trained convolutional neural network (CNN) by utilizing the experience replay technique and stochastic gradient descent (SGD). In addition, the Nash equilibrium and existence conditions of the proposed two schemes are theoretically deduced for the modeled MEC game against smart attacks. Compared with the traditional Q-learning algorithm, the average utility and secrecy capacity of the SU can be improved by the proposed DPESS and DESS schemes. Numerical simulations are also presented to verify the better performance of our proposals in terms of efficiency and security, including the higher convergence speed of the DESS strategy.

[1]  Huaguang Zhang,et al.  Fault-Tolerant Controller Design for a Class of Nonlinear MIMO Discrete-Time Systems via Online Reinforcement Learning Algorithm , 2016, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[2]  Wendi B. Heinzelman,et al.  Reinforcement learning in MIMO wireless networks with energy harvesting , 2017, 2017 IEEE International Conference on Communications (ICC).

[3]  Weihua Zhuang,et al.  Game Theoretic Study on Channel-Based Authentication in MIMO Systems , 2017, IEEE Transactions on Vehicular Technology.

[4]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[5]  Weihua Zhuang,et al.  Channel-Based Authentication Game in MIMO Systems , 2016, 2016 IEEE Global Communications Conference (GLOBECOM).

[6]  Xu Chen,et al.  Decentralized Computation Offloading Game for Mobile Cloud Computing , 2014, IEEE Transactions on Parallel and Distributed Systems.

[7]  Peter Dayan,et al.  Technical Note: Q-Learning , 2004, Machine Learning.

[8]  Lin Gao,et al.  Cooperative Spectrum Sharing: A Contract-Based Approach , 2014, IEEE Transactions on Mobile Computing.

[9]  Shuguang Cui,et al.  Joint offloading and computing optimization in wireless powered mobile-edge computing systems , 2017, 2017 IEEE International Conference on Communications (ICC).

[10]  Weihua Zhuang,et al.  User-Centric View of Unmanned Aerial Vehicle Transmission Against Smart Attacks , 2018, IEEE Transactions on Vehicular Technology.

[11]  Liang Xiao,et al.  Mobile cloud offloading for malware detections with learning , 2015, 2015 IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS).

[12]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[13]  Xiaojiang Du,et al.  Reinforcement Learning Based Mobile Offloading for Cloud-Based Malware Detection , 2017, GLOBECOM 2017 - 2017 IEEE Global Communications Conference.

[14]  Shaolei Ren,et al.  Online Learning for Offloading and Autoscaling in Energy Harvesting Mobile Edge Computing , 2017, IEEE Transactions on Cognitive Communications and Networking.

[15]  K. B. Letaief,et al.  A Survey on Mobile Edge Computing: The Communication Perspective , 2017, IEEE Communications Surveys & Tutorials.

[16]  Mohamed Ayoub Messous,et al.  Computation offloading game for an UAV network in mobile edge computing , 2017, 2017 IEEE International Conference on Communications (ICC).

[17]  H. Vincent Poor,et al.  A Mobile Offloading Game Against Smart Attacks , 2016, IEEE Access.

[18]  H. Vincent Poor,et al.  Mobile Crowdsensing Games in Vehicular Networks , 2017, IEEE Transactions on Vehicular Technology.

[19]  H. Vincent Poor,et al.  Game theoretic study of protecting MIMO transmissions against smart attacks , 2017, 2017 IEEE International Conference on Communications (ICC).

[20]  H. Vincent Poor,et al.  Mobile offloading game against smart attacks , 2016, 2016 IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS).

[21]  Shaocheng Tong,et al.  Reinforcement Learning Design-Based Adaptive Tracking Control With Less Learning Parameters for Nonlinear Discrete-Time MIMO Systems , 2015, IEEE Transactions on Neural Networks and Learning Systems.

[22]  Liang Xiao,et al.  Learning-Based Privacy-Aware Offloading for Healthcare IoT With Energy Harvesting , 2019, IEEE Internet of Things Journal.

[23]  H. Vincent Poor,et al.  A Secure Mobile Crowdsensing Game With Deep Reinforcement Learning , 2018, IEEE Transactions on Information Forensics and Security.

[24]  David Hutchison,et al.  The Extended Cloud: Review and Analysis of Mobile Edge Computing and Fog From a Security and Resilience Perspective , 2017, IEEE Journal on Selected Areas in Communications.

[25]  Andrew W. Moore,et al.  Prioritized Sweeping: Reinforcement Learning with Less Data and Less Time , 1993, Machine Learning.