Energy-aware optimization of UAV base stations placement via decentralized multi-agent Q-learning

Unmanned aerial vehicles serving as aerial base stations (UAV-BSs) can be deployed to provide wireless connectivity to ground devices in events of increased network demand, pointsof-failure in existing infrastructure, or disasters. However, it is challenging to conserve the energy of UAVs during prolonged coverage tasks, considering their limited on-board battery capacity. Reinforcement learning-based (RL) approaches have been previously used to improve energy utilization of multiple UAVs, however, a central cloud controller is assumed to have complete knowledge of the end-devices’ locations, i.e., the controller periodically scans and sends updates for UAV decision-making. This assumption is impractical in dynamic network environments with UAVs serving mobile ground devices. To address this problem, we propose a decentralized Q-learning approach, where each UAVBS is equipped with an autonomous agent that maximizes the connectivity of mobile ground devices while improving its energy utilization. Experimental results show that the proposed design significantly outperforms the centralized approaches in jointly maximizing the number of connected ground devices and the energy utilization of the UAV-BSs.

[1]  Qingqing Wu,et al.  Joint Trajectory and Communication Design for Multi-UAV Enabled Wireless Networks , 2017, IEEE Transactions on Wireless Communications.

[2]  Arumugam Nallanathan,et al.  Multi-Agent Reinforcement Learning-Based Resource Allocation for UAV Networks , 2018, IEEE Transactions on Wireless Communications.

[3]  Wei Xu,et al.  Energy Efficient UAV Communication With Energy Harvesting , 2020, IEEE Transactions on Vehicular Technology.

[4]  Tracy Camp,et al.  A survey of mobility models for ad hoc network research , 2002, Wirel. Commun. Mob. Comput..

[5]  Chi Harold Liu,et al.  Distributed Energy-Efficient Multi-UAV Navigation for Long-Term Communication Coverage by Deep Reinforcement Learning , 2020, IEEE Transactions on Mobile Computing.

[6]  Chi Harold Liu,et al.  Energy-Efficient UAV Control for Effective and Fair Communication Coverage: A Deep Reinforcement Learning Approach , 2018, IEEE Journal on Selected Areas in Communications.

[7]  Jie Xu,et al.  Energy Minimization for Wireless Communication With Rotary-Wing UAV , 2018, IEEE Transactions on Wireless Communications.

[8]  Walid Saad,et al.  Mobile Unmanned Aerial Vehicles (UAVs) for Energy-Efficient Internet of Things Communications , 2017, IEEE Transactions on Wireless Communications.

[9]  Eduardo F. Morales,et al.  An Introduction to Reinforcement Learning , 2011 .

[10]  Luiz A. DaSilva,et al.  A Stochastic Model for UAV Networks Positioned Above Demand Hotspots in Urban Environments , 2018, IEEE Transactions on Vehicular Technology.

[11]  Luiz A. DaSilva,et al.  Coverage Analysis for Low-Altitude UAV Networks in Urban Environments , 2017, GLOBECOM 2017 - 2017 IEEE Global Communications Conference.

[12]  Xiao Liu,et al.  Reinforcement Learning in Multiple-UAV Networks: Deployment and Movement Design , 2019, IEEE Transactions on Vehicular Technology.

[13]  Jingjing Yao,et al.  QoS-Aware Power Control in Internet of Drones for Data Collection Service , 2019, IEEE Transactions on Vehicular Technology.

[14]  Karl Tuyls,et al.  An Overview of Cooperative and Competitive Multiagent Learning , 2005, LAMAS.

[15]  Inkyu Lee,et al.  UAV-Aided Wireless Communication Designs With Propulsion Energy Limitations , 2018, IEEE Transactions on Vehicular Technology.

[16]  Yongxin Liu,et al.  Reinforcement Learning Optimized Throughput for 5G Enhanced Swarm UAS Networking , 2021, ICC 2021 - IEEE International Conference on Communications.

[17]  Ming Tan,et al.  Multi-Agent Reinforcement Learning: Independent versus Cooperative Agents , 1997, ICML.

[18]  Halim Yanikomeroglu,et al.  Efficient 3-D placement of an aerial base station in next generation cellular networks , 2016, 2016 IEEE International Conference on Communications (ICC).

[19]  Sajal K. Das,et al.  Coverage, Connectivity, and Fault Tolerance Measures of Wireless Sensor Networks , 2006, SSS.

[20]  Ivana Dusparic,et al.  REQIBA: Regression and Deep Q-Learning for Intelligent UAV Cellular User to Base Station Association , 2020, IEEE Transactions on Vehicular Technology.

[21]  Luiz A. DaSilva,et al.  Deployment of UAV-mounted access points according to spatial user locations in two-tier cellular networks , 2016, 2016 Wireless Days (WD).

[22]  Hyun-Rok Cha,et al.  Practical Endurance Estimation for Minimizing Energy Consumption of Multirotor Unmanned Aerial Vehicles , 2018, Energies.

[23]  Maxime Guériau,et al.  An RL-based Approach to Improve Communication Performance and Energy Utilization in Fog-based IoT , 2019, 2019 International Conference on Wireless and Mobile Computing, Networking and Communications (WiMob).

[24]  Bart De Schutter,et al.  A Comprehensive Survey of Multiagent Reinforcement Learning , 2008, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).