Learning to Entangle Radio Resources in Vehicular Communications: An Oblivious Game-Theoretic Perspective

In this paper, we investigate non-cooperative radio resource management in a vehicle-to-vehicle communication network. The technical challenges lie in high-vehicle mobility and data traffic variations. Over the discrete scheduling slots, each vehicle user equipment (VUE)-pair competes with other VUE-pairs in the coverage of a road side unit (RSU) for the limited frequency to transmit queued data packets, aiming to optimize the expected long-term performance. The frequency allocation at the beginning of each slot at the RSU is regulated by a sealed second-price auction. Such interactions among VUE-pairs are modeled as a stochastic game with a semi-continuous global network state space. By defining a partitioned control policy, we transform the original game into an equivalent stochastic game with a global queue state space of finite size. We adopt an oblivious equilibrium (OE) to approximate the Markov perfect equilibrium, which characterizes the optimal solution to the equivalent game. The OE solution is theoretically proven to be with an asymptotic Markov equilibrium property. Due to the lack of a priori knowledge of network dynamics, we derive an online algorithm to learn the OE solution. Numerical simulations validate the theoretical analysis and show the effectiveness of the proposed online learning algorithm.

[1]  Ulrike von Luxburg,et al.  A tutorial on spectral clustering , 2007, Stat. Comput..

[2]  William Vickrey,et al.  Counterspeculation, Auctions, And Competitive Sealed Tenders , 1961 .

[3]  Henk Wymeersch,et al.  Fine-Grained vs. Average Reliability for V2V Communications around Intersections , 2017, 2017 IEEE Globecom Workshops (GC Wkshps).

[4]  Yu Zhang,et al.  A PCA-based modeling method for wireless MIMO channel , 2017, 2017 IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS).

[5]  Ahmed Helmy,et al.  IMPORTANT: a framework to systematically analyze the Impact of Mobility on Performance of Routing Protocols for Adhoc Networks , 2003, IEEE INFOCOM 2003. Twenty-second Annual Joint Conference of the IEEE Computer and Communications Societies (IEEE Cat. No.03CH37428).

[6]  A. Lerner COUNTERSPECULATION, AUCTIONS, AND COMPETITIVE SEALED TENDERS , 2012 .

[7]  Mehdi Bennis,et al.  Performance Optimization in Mobile-Edge Computing via Deep Reinforcement Learning , 2018, 2018 IEEE 88th Vehicular Technology Conference (VTC-Fall).

[8]  Mehrdad Dianati,et al.  A Survey of the State-of-the-Art Localization Techniques and Their Potentials for Autonomous Vehicle Applications , 2018, IEEE Internet of Things Journal.

[9]  Mihaela van der Schaar,et al.  Learning to Compete for Resources in Wireless Stochastic Games , 2009, IEEE Transactions on Vehicular Technology.

[10]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[11]  Vincent K. N. Lau,et al.  Queue-Aware Dynamic Clustering and Power Allocation for Network MIMO Systems via Distributed Stochastic Learning , 2010, IEEE Transactions on Signal Processing.

[12]  Abhijeet Bhorkar,et al.  An on-line learning algorithm for energy efficient delay constrained scheduling over a fading channel , 2008, IEEE Journal on Selected Areas in Communications.

[13]  Erik G. Ström,et al.  Radio Resource Management for D2D-Based V2V Communication , 2016, IEEE Transactions on Vehicular Technology.

[14]  C. L. Benkard,et al.  Markov Perfect Industry Dynamics with Many Firms , 2005 .

[15]  Jianhua Zhang,et al.  Clustering in wireless propagation channel with a statistics-based framework , 2018, 2018 IEEE Wireless Communications and Networking Conference (WCNC).

[16]  Zhu Han,et al.  Wireless Resource Scheduling in Virtualized Radio Access Networks Using Stochastic Learning , 2018, IEEE Transactions on Mobile Computing.

[17]  Walid Saad,et al.  Distributed Federated Learning for Ultra-Reliable Low-Latency Vehicular Communications , 2018, IEEE Transactions on Communications.

[18]  Jianping Pan,et al.  On the Uplink MAC Performance of a Drive-Thru Internet , 2012, IEEE Transactions on Vehicular Technology.

[19]  Antonio Iera,et al.  5G Network Slicing for Vehicle-to-Everything Services , 2017, IEEE Wireless Communications.

[20]  Yusheng Ji,et al.  Cluster-Based Content Distribution Integrating LTE and IEEE 802.11p with Fuzzy Logic and Q-Learning , 2018, IEEE Computational Intelligence Magazine.

[21]  Pablo Pavón-Mariño,et al.  Distributed and Fair Beaconing Rate Adaptation for Congestion Control in Vehicular Networks , 2016, IEEE Transactions on Mobile Computing.

[22]  Zhigang Cao,et al.  Low Complexity Outage Optimal Distributed Channel Allocation for Vehicle-to-Vehicle Communications , 2011, IEEE Journal on Selected Areas in Communications.

[23]  Yiqing Zhou,et al.  Heterogeneous Vehicular Networking: A Survey on Architecture, Challenges, and Solutions , 2015, IEEE Communications Surveys & Tutorials.

[24]  Vincent K. N. Lau,et al.  Delay-Aware Two-Hop Cooperative Relay Communications via Approximate MDP and Stochastic Learning , 2013, IEEE Transactions on Information Theory.

[25]  H. T. Kung,et al.  Competing Mobile Network Game: Embracing antijamming and jamming strategies with reinforcement learning , 2013, 2013 IEEE Conference on Communications and Network Security (CNS).

[26]  Antonella Molinaro,et al.  Information-centric networking for connected vehicles: a survey and future perspectives , 2016, IEEE Communications Magazine.

[27]  Victor C. M. Leung,et al.  Delay-Optimal Virtualized Radio Resource Scheduling in Software-Defined Vehicular Networks via Stochastic Learning , 2016, IEEE Transactions on Vehicular Technology.

[28]  Mehdi Bennis,et al.  Ultra-Reliable and Low-Latency Vehicular Transmission: An Extreme Value Theory Approach , 2018, IEEE Communications Letters.

[29]  Benjamin Van Roy,et al.  Oblivious Equilibrium: A Mean Field Approximation for Large-Scale Dynamic Games , 2005, NIPS.

[30]  Xue Liu,et al.  LORA: Loss Differentiation Rate Adaptation Scheme for Vehicle-to-Vehicle Safety Communications , 2017, IEEE Transactions on Vehicular Technology.

[31]  Mehdi Bennis,et al.  Multi-Tenant Cross-Slice Resource Orchestration: A Deep Reinforcement Learning Approach , 2018, IEEE Journal on Selected Areas in Communications.

[32]  Mihaela van der Schaar,et al.  Joint Physical-Layer and System-Level Power Management for Delay-Sensitive Wireless Communications , 2013, IEEE Transactions on Mobile Computing.

[33]  Andrea J. Goldsmith,et al.  On oblivious equilibrium in large population stochastic games , 2010, 49th IEEE Conference on Decision and Control (CDC).

[34]  Mehdi Bennis,et al.  Optimized Computation Offloading Performance in Virtual Edge Computing Systems Via Deep Reinforcement Learning , 2018, IEEE Internet of Things Journal.

[35]  Yu Zhang,et al.  Clustering Analysis in the Wireless Propagation Channel with a Variational Gaussian Mixture Model , 2020, IEEE Transactions on Big Data.

[36]  Albert Banchs,et al.  Mobile network architecture evolution toward 5G , 2016, IEEE Communications Magazine.

[37]  T. E. S. Raghavan,et al.  Existence of Stationary Correlated Equilibria with Symmetric Information for Discounted Stochastic Games , 1992, Math. Oper. Res..

[38]  Mehdi Bennis,et al.  An Oblivious Game-Theoretic Approach for Wireless Scheduling in V2V Communications , 2017, GLOBECOM 2017 - 2017 IEEE Global Communications Conference.

[39]  Dimitri P. Bertsekas,et al.  Dynamic Programming: Deterministic and Stochastic Models , 1987 .