Robust Deep Reinforcement Learning for Security and Safety in Autonomous Vehicle Systems

The dependence of autonomous vehicles (AVs) on sensors and communication links exposes them to cyber-physical (CP) attacks by adversaries that seek to take control of the AVs by manipulating their data. In this paper, the state estimation process for monitoring AV dynamics, in presence of CP attacks, is analyzed and a novel adversarial deep reinforcement learning (RL) algorithm is proposed to maximize the robustness of AV dynamics control to CP attacks. The attacker's action and the AV's reaction to CP attacks are studied in a game-theoretic framework. In the formulated game, the attacker seeks to inject faulty data to AV sensor readings so as to manipulate the inter-vehicle optimal safe spacing and potentially increase the risk of AV accidents or reduce the vehicle flow on the roads. Meanwhile, the AV, acting as a defender, seeks to minimize the deviations of spacing so as to ensure robustness to the attacker's actions. Since the AV has no information about the attacker's action and due to the infinite possibilities for data value manipulations, each player uses long short term memory (LSTM) blocks to learn the expected spacing deviation resulting from its own action and feeds this deviation to a reinforcement learning (RL) algorithm. Then, the attacker's RL algorithm chooses the action which maximizes the spacing deviation, while the AV's RL algorithm seeks to find the optimal action that minimizes such deviation. Simulation results show that the proposed adversarial deep RL algorithm can improve the robustness of the AV dynamics control as it minimizes the intra-AV spacing deviation.

[1]  Naresh K. Sinha,et al.  Modern Control Systems , 1981, IEEE Transactions on Systems, Man, and Cybernetics.

[2]  Walid Saad,et al.  Unmanned Aerial Vehicle With Underlaid Device-to-Device Communications: Performance and Tradeoffs , 2015, IEEE Transactions on Wireless Communications.

[3]  Zhu Han,et al.  Game Theory in Wireless and Communication Networks: Theory, Models, and Applications , 2011 .

[4]  T. Başar,et al.  Dynamic Noncooperative Game Theory , 1982 .

[5]  Yves Deswarte,et al.  Survey on security threats and protection mechanisms in embedded automotive networks , 2013, 2013 43rd Annual IEEE/IFIP Conference on Dependable Systems and Networks Workshop (DSN-W).

[6]  David Silver,et al.  Deep Reinforcement Learning from Self-Play in Imperfect-Information Games , 2016, ArXiv.

[7]  Stefania Santini,et al.  A collaborative approach for improving the security of vehicular scenarios: The case of platooning , 2018, Comput. Commun..

[8]  Walid Saad,et al.  Colonel Blotto Game for Secure State Estimation in Interdependent Critical Infrastructure , 2017, ArXiv.

[9]  Geoffrey E. Hinton,et al.  Speech recognition with deep recurrent neural networks , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[10]  Walid Saad,et al.  Deep Learning for Reliable Mobile Edge Analytics in Intelligent Transportation Systems: An Overview , 2017, IEEE Vehicular Technology Magazine.

[11]  Panagiotis Papadimitratos,et al.  On the Performance of Secure Vehicular Communication Systems , 2011, IEEE Transactions on Dependable and Secure Computing.

[12]  Dipak Ghosal,et al.  Security vulnerabilities of connected vehicle streams and their impact on cooperative driving , 2015, IEEE Communications Magazine.

[13]  Walid Saad,et al.  Deep Learning for Signal Authentication and Security in Massive Internet-of-Things Systems , 2018, IEEE Transactions on Communications.

[14]  Walid Saad,et al.  Machine Learning for Wireless Networks with Artificial Intelligence: A Tutorial on Neural Networks , 2017, ArXiv.

[15]  Walid Saad,et al.  Joint Communication and Control for Wireless Autonomous Vehicular Platoon Systems , 2018, IEEE Transactions on Communications.

[16]  Panagiotis Papadimitratos,et al.  Secure vehicular communication systems: implementation, performance, and research challenges , 2008, IEEE Communications Magazine.

[17]  Dong Hoon Lee,et al.  A Practical Wireless Attack on the Connected Car and Security Protocol for In-Vehicle CAN , 2015, IEEE Transactions on Intelligent Transportation Systems.

[18]  Ryan M. Gerdes,et al.  A data trust framework for VANETs enabling false data detection and secure vehicle tracking , 2017, 2017 IEEE Conference on Communications and Network Security (CNS).

[19]  Zhu Han,et al.  Game Theory in Wireless and Communication Networks , 2008 .

[20]  Xin Zhang,et al.  VANET alert endorsement using multi-source filters , 2010, VANET '10.

[21]  Walid Saad,et al.  A colonel blotto game for interdependence-aware cyber-physical systems security in smart cities , 2017, SCOPE@CPSWeek.

[22]  Tomas Olovsson,et al.  Security aspects of the in-vehicle network in the connected car , 2011, 2011 IEEE Intelligent Vehicles Symposium (IV).

[23]  Ursula Challita,et al.  Artificial Neural Networks-Based Machine Learning for Wireless Networks: A Tutorial , 2017, IEEE Communications Surveys & Tutorials.

[24]  Mike McDonald,et al.  Car-following: a historical review , 1999 .

[25]  Liam Kilmartin,et al.  Intra-Vehicle Networks: A Review , 2015, IEEE Transactions on Intelligent Transportation Systems.

[26]  Theodore Bohn,et al.  Security concerns of a plug-in vehicle , 2012, 2012 IEEE PES Innovative Smart Grid Technologies (ISGT).