Learning-Based Energy-Efficient Data Collection by Unmanned Vehicles in Smart Cities

Mobile crowdsourcing (MCS) is now an important source of information for smart cities, especially with the help of unmanned aerial vehicles (UAVs) and driverless cars. They are equipped with different kinds of high-precision sensors, and can be scheduled/controlled completely during data collection, which will make MCS system more robust. However, they are limited to energy constraint, especially for long-term, long-distance sensing tasks, and cities are almost too crowded to set stationary charging station. Towards this end, in this paper we propose to leverage emerging deep reinforcement learning (DRL) techniques for enabling model-free unmanned vehicles control, and present a novel and highly effective control framework, called “DRL-RVC.” It utilizes the powerful convolutional neural network for feature extraction of the necessary information (including sample distribution, traffic flow, etc.), then makes decisions under the guidance of the deep Q network. That is, UAVs will cruise in the city without control and collect most required data in the sensing region, while mobile unmanned charging station will reach the charging point in the shortest possible time. Finally, we validate and evaluate the proposed framework via extensive simulations based on a real dataset in Rome. Extensive simulation results well justify the effectiveness and robustness of our approach.

[1]  Yuval Tassa,et al.  Continuous control with deep reinforcement learning , 2015, ICLR.

[2]  Milan Simic,et al.  Investigation in Wireless Power Transmission for UAV Charging , 2015, KES.

[3]  Tom Schaul,et al.  Prioritized Experience Replay , 2015, ICLR.

[4]  Victor C. M. Leung,et al.  CAP: community activity prediction based on big data analysis , 2014, IEEE Network.

[5]  Yan Zhang,et al.  Enabling Localized Peer-to-Peer Electricity Trading Among Plug-in Hybrid Electric Vehicles Using Consortium Blockchains , 2017, IEEE Transactions on Industrial Informatics.

[6]  Chris Watkins,et al.  Learning from delayed rewards , 1989 .

[7]  Honglak Lee,et al.  Control of Memory, Active Perception, and Action in Minecraft , 2016, ICML.

[8]  Yishay Mansour,et al.  Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.

[9]  Jie Wu,et al.  Toward QoI and Energy Efficiency in Participatory Crowdsourcing , 2015, IEEE Transactions on Vehicular Technology.

[10]  Chi Harold Liu,et al.  Unsupervised posture detection by smartphone accelerometer , 2013 .

[11]  Suman Nath ACE: Exploiting Correlation for Energy-Efficient and Continuous Context Sensing , 2013, IEEE Trans. Mob. Comput..

[12]  Jürgen Schmidhuber,et al.  A Machine Learning Approach to Visual Perception of Forest Trails for Mobile Robots , 2016, IEEE Robotics and Automation Letters.

[13]  Andrew W. Moore,et al.  Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..

[14]  Mohsen Guizani,et al.  Home M2M networks: Architectures, standards, and QoS improvement , 2011, IEEE Communications Magazine.

[15]  Trevor Darrell,et al.  Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[16]  Fei-Fei Li,et al.  Deep visual-semantic alignments for generating image descriptions , 2015, CVPR.

[17]  Kin K. Leung,et al.  Energy-Aware Participant Selection for Smartphone-Enabled Mobile Crowd Sensing , 2017, IEEE Systems Journal.

[18]  Kin K. Leung,et al.  Dynamic Control of Data Ferries under Partial Observations , 2010, 2010 IEEE Wireless Communication and Networking Conference.

[19]  Sergey Levine,et al.  High-Dimensional Continuous Control Using Generalized Advantage Estimation , 2015, ICLR.

[20]  Antonio Corradi,et al.  The participact mobile crowd sensing living lab: The testbed for smart cities , 2014, IEEE Communications Magazine.

[21]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[22]  David Silver,et al.  Deep Reinforcement Learning with Double Q-Learning , 2015, AAAI.

[23]  Kilian Q. Weinberger,et al.  Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[24]  Martín Abadi,et al.  TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems , 2016, ArXiv.

[25]  Mark A. Richardson,et al.  A Low-cost Vision Based Navigation System for Small Size Unmanned Aerial Vehicle Applications , 2013 .

[26]  Yan Zhang,et al.  Optimal Incentive Design for Cloud-Enabled Multimedia Crowdsourcing , 2016, IEEE Transactions on Multimedia.

[27]  Jan Peters,et al.  A Survey on Policy Search for Robotics , 2013, Found. Trends Robotics.

[28]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[29]  Victor C. M. Leung,et al.  SOVCAN: Safety-Oriented Vehicular Controller Area Network , 2017, IEEE Communications Magazine.

[30]  Quanyan Zhu,et al.  Dependable Demand Response Management in the Smart Grid: A Stackelberg Game Approach , 2013, IEEE Transactions on Smart Grid.

[31]  Victor Pankratius,et al.  Mobile crowd sensing in space weather monitoring: the mahali project , 2014, IEEE Communications Magazine.

[32]  Yan Zhang,et al.  Software Defined Machine-to-Machine Communication for Smart Energy Management , 2017, IEEE Communications Magazine.

[33]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.