A Non-Stochastic Learning Approach to Energy Efficient Mobility Management

Energy efficient mobility management is an important problem in modern wireless networks with heterogeneous cell sizes and increased nodes densities. We show that optimization-based mobility protocols cannot achieve long-term optimal energy consumption, particularly for ultra-dense networks (UDNs). To address the complex dynamics of UDN, we propose a non-stochastic online-learning approach, which does not make any assumption on the statistical behavior of the small base station (SBS) activities. In addition, we introduce handover cost to the overall energy consumption, which forces the resulting solution to explicitly minimize frequent handovers. The proposed batched randomization with exponential weighting (BREW) algorithm relies on batching to explore in bulk, and hence reduces unnecessary handovers. We prove that the regret of BREW is sublinear in time, thus guaranteeing its convergence to the optimal SBS selection. We further study the robustness of the BREW algorithm to delayed or missing feedback. Moreover, we study the setting where SBSs can be dynamically turned ON and OFF. We prove that sublinear regret is impossible with respect to arbitrary SBS ON/OFF, and then develop a novel learning strategy, called ranking expert (RE), that simultaneously takes into account the handover cost and the availability of SBS. To address the high complexity of RE, we propose a contextual ranking expert (CRE) algorithm that only assigns experts in a given context. Rigorous regret bounds are proved for both RE and CRE with respect to the best expert. Simulations show that not only do the proposed mobility algorithms greatly reduce the system energy consumption, but they are also robust to various dynamics which are common in practical ultra-dense wireless networks.

[1]  Robert W. Heath,et al.  Five disruptive technology directions for 5G , 2013, IEEE Communications Magazine.

[2]  Afef Feki,et al.  Joint interference management and handover optimization in LTE small cells network , 2012, 2012 IEEE International Conference on Communications (ICC).

[3]  Ismail Güvenç,et al.  Context-aware mobility management in HetNets: A reinforcement learning approach , 2015, 2015 IEEE Wireless Communications and Networking Conference (WCNC).

[4]  Robert D. Kleinberg,et al.  Regret bounds for sleeping experts and bandits , 2010, Machine Learning.

[5]  Neri Merhav,et al.  On sequential strategies for loss functions with memory , 2002, IEEE Trans. Inf. Theory.

[6]  Tracy Camp,et al.  A survey of mobility models for ad hoc network research , 2002, Wirel. Commun. Mob. Comput..

[7]  Fabrizio Granelli,et al.  A Handover Policy for Energy Efficient Network Connectivity through Proportionally Fair Access , 2014 .

[8]  T. L. Lai Andherbertrobbins Asymptotically Efficient Adaptive Allocation Rules , 2022 .

[9]  Christian Vitale,et al.  Energy-efficient user association in extremely dense small cell networks , 2014, 2014 European Conference on Networks and Communications (EuCNC).

[10]  Tony Q. S. Quek,et al.  Small Cell Networks: Deployment, PHY Techniques, and Resource Management , 2013 .

[11]  Filip Radlinski,et al.  Mortal Multi-Armed Bandits , 2008, NIPS.

[12]  Sébastien Bubeck,et al.  Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems , 2012, Found. Trends Mach. Learn..

[13]  Mingyan Liu,et al.  Online Learning of Rested and Restless Bandits , 2011, IEEE Transactions on Information Theory.

[14]  Yuval Peres,et al.  Bandits with switching costs: T2/3 regret , 2013, STOC.

[15]  Stefania Sesia,et al.  LTE - The UMTS Long Term Evolution, Second Edition , 2011 .

[16]  Luis Alonso,et al.  Energy-efficient context-aware user association for outdoor small cell heterogeneous networks , 2014, 2014 IEEE International Conference on Communications (ICC).

[17]  Jeffrey G. Andrews,et al.  User Association for Load Balancing in Heterogeneous Cellular Networks , 2012, IEEE Transactions on Wireless Communications.

[18]  Gábor Lugosi,et al.  Prediction, learning, and games , 2006 .

[19]  Lazaros F. Merakos,et al.  Energy-efficient and interference-aware handover decision for the LTE-Advanced femtocell network , 2013, 2013 IEEE International Conference on Communications (ICC).

[20]  Rami Puzis,et al.  Volatile Multi-Armed Bandits for Guaranteed Targeted Social Crawling , 2013, AAAI.

[21]  Peter Auer,et al.  Finite-time Analysis of the Multiarmed Bandit Problem , 2002, Machine Learning.

[22]  Ambuj Tewari,et al.  Online Bandit Learning against an Adaptive Adversary: from Regret to Policy Regret , 2012, ICML.

[23]  Peter Auer,et al.  The Nonstochastic Multiarmed Bandit Problem , 2002, SIAM J. Comput..

[24]  Jeffrey G. Andrews,et al.  An overview of load balancing in hetnets: old myths and open problems , 2013, IEEE Wireless Communications.

[25]  Sébastien Bubeck Jeux de bandits et fondations du clustering , 2010 .

[26]  Stefania Sesia,et al.  LTE - The UMTS Long Term Evolution , 2009 .