Multi-Task-Oriented Vehicular Crowdsensing: A Deep Learning Approach

With the popularity of drones and driverless cars, vehicular crowdsensing (VCS) becomes increasingly widely-used by taking advantage of their high-precision sensors and durability in harsh environments. Since abrupt sensing tasks usually cannot be prepared beforehand, we need a generic control logic fit-for-use all tasks which are similar in nature, but different in their own settings like Point-of-Interest (PoI) distributions. The objectives include to simultaneously maximize the data collection amount, geographic fairness, and minimize the energy consumption of all vehicles for all tasks, which usually cannot be explicitly expressed in a closed-form equation, thus not tractable as an optimization problem. In this paper, we propose a deep reinforcement learning (DRL)-based centralized control, distributed execution framework for multi-task-oriented VCS, called "DRL-MTVCS". It includes an asynchronous architecture with spatiotemporal state information modeling, multi-task-oriented value estimates by adaptive normalization, and auxiliary vehicle action exploration by pixel control. We compare with three baselines, and results show that DRL-MTVCS outperforms all others in terms of energy efficiency when varying different numbers of tasks, vehicles, charging stations and sensing range.

[1]  Daqing Zhang,et al.  iCrowd: Near-Optimal Task Allocation for Piggyback Crowdsensing , 2016, IEEE Transactions on Mobile Computing.

[2]  Salimur Choudhury,et al.  Improved Recruitment Algorithms for Vehicular Crowdsensing Networks , 2019, IEEE Transactions on Vehicular Technology.

[3]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[4]  H. Vincent Poor,et al.  Mobile Crowdsensing Games in Vehicular Networks , 2017, IEEE Transactions on Vehicular Technology.

[5]  Kin K. Leung,et al.  Dynamic Control of Data Ferries under Partial Observations , 2010, 2010 IEEE Wireless Communication and Networking Conference.

[6]  Dzmitry Kliazovich,et al.  A Survey on Mobile Crowdsensing Systems: Challenges, Solutions, and Opportunities , 2019, IEEE Communications Surveys & Tutorials.

[7]  Xiumin Wang,et al.  Mobility-Aware Participant Recruitment for Vehicle-Based Mobile Crowdsensing , 2018, IEEE Transactions on Vehicular Technology.

[8]  Tom Schaul,et al.  Reinforcement Learning with Unsupervised Auxiliary Tasks , 2016, ICLR.

[9]  Hui Gao,et al.  Online Quality-Aware Incentive Mechanism for Mobile Crowd Sensing with Extra Bonus , 2019, IEEE Transactions on Mobile Computing.

[10]  Chi Harold Liu,et al.  Hybrid Vehicular Crowdsourcing With Driverless Cars: Challenges and a Solution , 2018, Computer.

[11]  Raj Jain,et al.  A Quantitative Measure Of Fairness And Discrimination For Resource Allocation In Shared Computer Systems , 1998, ArXiv.

[12]  Chi Harold Liu,et al.  Free Market of Multi-Leader Multi-Follower Mobile Crowdsensing: An Incentive Mechanism Design by Deep Reinforcement Learning , 2020, IEEE Transactions on Mobile Computing.

[13]  Chi Harold Liu,et al.  Energy-Efficient Distributed Mobile Crowd Sensing: A Deep Learning Approach , 2019, IEEE Journal on Selected Areas in Communications.

[14]  E. Ionides Truncated Importance Sampling , 2008 .

[15]  Wojciech Czarnecki,et al.  Multi-task Deep Reinforcement Learning with PopArt , 2018, AAAI.

[16]  Mohsen Guizani,et al.  When Mobile Crowd Sensing Meets UAV: Energy-Efficient Task Assignment and Route Planning , 2018, IEEE Transactions on Communications.

[17]  Daqing Zhang,et al.  Near-Optimal Incentive Allocation for Piggyback Crowdsensing , 2017, IEEE Communications Magazine.

[18]  Tom Schaul,et al.  Rainbow: Combining Improvements in Deep Reinforcement Learning , 2017, AAAI.

[19]  Alex Graves,et al.  Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.

[20]  Shane Legg,et al.  IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures , 2018, ICML.

[21]  Merkourios Karaliopoulos,et al.  User recruitment for mobile crowdsensing over opportunistic networks , 2015, 2015 IEEE Conference on Computer Communications (INFOCOM).

[22]  Tom Schaul,et al.  Dueling Network Architectures for Deep Reinforcement Learning , 2015, ICML.

[23]  Shilin Wen,et al.  Blockchain-Enabled Data Collection and Sharing for Industrial IoT With Deep Reinforcement Learning , 2019, IEEE Transactions on Industrial Informatics.

[24]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[25]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[26]  Xiaoying Gan,et al.  Dynamic Task Assignment in Crowdsensing with Location Awareness and Location Diversity , 2018, IEEE INFOCOM 2018 - IEEE Conference on Computer Communications.

[27]  Yunhao Liu,et al.  Incentives for Mobile Crowd Sensing: A Survey , 2016, IEEE Communications Surveys & Tutorials.

[28]  Guihai Chen,et al.  Data-Oriented Mobile Crowdsensing: A Comprehensive Survey , 2019, IEEE Communications Surveys & Tutorials.

[29]  Jon Crowcroft,et al.  Distributed and Energy-Efficient Mobile Crowdsensing with Charging Stations by Deep Reinforcement Learning , 2021, IEEE Transactions on Mobile Computing.