Deep Reinforcement Learning based Distributed Resource Allocation for V2V Broadcasting

In this article, we exploit deep reinforcement learning for joint resource allocation and scheduling in vehicle-to-vehicle (V2V) broadcast communications. Each vehicle, considered as an autonomous agent, makes its decisions to find the messages and spectrum for transmission based on its local observations without requiring or having to wait for global information. From the simulation results, each vehicle can effectively learn how to ensure the stringent latency constraints on V2V links while minimizing the interference to vehicle-to-infrastructure (V2I) links.

[1]  Zhigang Cao,et al.  Low Complexity Outage Optimal Distributed Channel Allocation for Vehicle-to-Vehicle Communications , 2011, IEEE Journal on Selected Areas in Communications.

[2]  Yu-Chee Tseng,et al.  The Broadcast Storm Problem in a Mobile Ad Hoc Network , 1999, Wirel. Networks.

[3]  Geoffrey Ye Li,et al.  Toward Intelligent Vehicular Networks: A Machine Learning Framework , 2018, IEEE Internet of Things Journal.

[4]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[5]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[6]  Ozan K. Tonguz,et al.  On the Broadcast Storm Problem in Ad hoc Wireless Networks , 2006, 2006 3rd International Conference on Broadband Communications, Networks and Systems.

[7]  Biing-Hwang Juang,et al.  Recurrent deep neural networks for robust speech recognition , 2014, 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[8]  Geoffrey Ye Li,et al.  Device-to-device communications in cellular networks , 2016, IEEE Communications Magazine.

[9]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[10]  Geoffrey Ye Li,et al.  Vehicular Communications: A Network Layer Perspective , 2017, IEEE Transactions on Vehicular Technology.

[11]  Srikanth Kandula,et al.  Resource Management with Deep Reinforcement Learning , 2016, HotNets.

[12]  Geoffrey Ye Li,et al.  Deep Reinforcement Learning for Resource Allocation in V2V Communications , 2017, 2018 IEEE International Conference on Communications (ICC).

[13]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[14]  Geoffrey Ye Li,et al.  Power of Deep Learning for Channel Estimation and Signal Detection in OFDM Systems , 2017, IEEE Wireless Communications Letters.

[15]  Erik G. Ström,et al.  Radio Resource Management for D2D-Based V2V Communication , 2016, IEEE Transactions on Vehicular Technology.

[16]  Geoffrey Ye Li,et al.  Resource Allocation for D2D-Enabled Vehicular Communications , 2017, IEEE Transactions on Communications.

[17]  Geoffrey Ye Li,et al.  Machine Learning for Vehicular Networks , 2017, ArXiv.

[18]  Philippe J. Sartori,et al.  LTE evolution for vehicle-to-everything services , 2016, IEEE Communications Magazine.

[19]  Demis Hassabis,et al.  Mastering the game of Go with deep neural networks and tree search , 2016, Nature.

[20]  Geoffrey Ye Li,et al.  Vehicular Communications: A Physical Layer Perspective , 2017, IEEE Transactions on Vehicular Technology.

[21]  Walid Saad,et al.  Dynamic Proximity-Aware Resource Allocation in Vehicle-to-Vehicle (V2V) Communications , 2016, 2016 IEEE Globecom Workshops (GC Wkshps).