Scalable Parallel Task Scheduling for Autonomous Driving Using Multi-Task Deep Reinforcement Learning

The Internet of Vehicles (IoV) as a promising application of Internet of Things (IoT) has played a significant role in autonomous driving, by connecting intelligent vehicles. Autonomous driving needs to process the mass environmental sensing data in coordination with surrounding vehicles, and makes an accurate driving judgment accordingly. Since the vehicles always have limited computing resources, processing these data in parallel with efficient task scheduling is one of the most important topics. Most current work focuses on formulating special scenarios and service requirements as optimization problems. However, the complicated and dynamic environment of vehicular computing is hard to model, predict and control, making those previous methods unscalable and unable to reflect the real scenario. In this paper, a Multi-task Deep reinforcement learning approach for scalable parallel Task Scheduling (MDTS) is firstly devised. For avoiding the curse of dimensionality when coping with complex parallel computing environments and jobs with diverse properties, we extend the action selection in Deep Reinforcement Learning (DRL) to a multi-task decision, where the output branches of multi- task learning are fine-matched to parallel scheduling tasks. Child tasks of a job are accordingly assigned to distributed nodes without any human knowledge while the resource competition among parallel tasks is leveraged through shared neural network layers. Moreover, we design an appropriate reward function to optimize multiple metrics simultaneously, with emphasis on specific scenarios. Extensive experiments show that the MDTS significantly increases the overall reward compared with least- connection scheduling and particle swarm optimization algorithm from −16.71, −0.67 to 2.93, respectively.

[1]  Wei Tan,et al.  Self-Adaptive Learning PSO-Based Deadline Constrained Task Scheduling for Hybrid IaaS Cloud , 2014, IEEE Transactions on Automation Science and Engineering.

[2]  Ji Li,et al.  DRL-cloud: Deep reinforcement learning-based resource provisioning and task scheduling for cloud service providers , 2018, 2018 23rd Asia and South Pacific Design Automation Conference (ASP-DAC).

[3]  Massoud Pedram,et al.  A Reinforcement Learning-Based Power Management Framework for Green Computing Data Centers , 2016, 2016 IEEE International Conference on Cloud Engineering (IC2E).

[4]  Nei Kato,et al.  An Intelligent Route Computation Approach Based on Real-Time Deep Learning Strategy for Software Defined Communication Systems , 2019, IEEE Transactions on Emerging Topics in Computing.

[5]  Rong Yu,et al.  Social Welfare Maximization in Container-Based Task Scheduling for Parked Vehicle Edge Computing , 2019, IEEE Communications Letters.

[6]  Ke Zhang,et al.  Mobile-Edge Computing for Vehicular Networks: A Promising Network Paradigm with Predictive Off-Loading , 2017, IEEE Veh. Technol. Mag..

[7]  Qinru Qiu,et al.  A Hierarchical Framework of Cloud Resource Allocation and Power Management Using Deep Reinforcement Learning , 2017, 2017 IEEE 37th International Conference on Distributed Computing Systems (ICDCS).

[8]  Alex Graves,et al.  Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.

[9]  Weihua Zhuang,et al.  Learning-Based Computation Offloading for IoT Devices With Energy Harvesting , 2017, IEEE Transactions on Vehicular Technology.

[10]  Jianping Pan,et al.  Joint Roadside Unit Deployment and Service Task Assignment for Internet of Vehicles (IoV) , 2019, IEEE Internet of Things Journal.

[11]  Nei Kato,et al.  A Novel Non-Supervised Deep-Learning-Based Network Traffic Control Method for Software Defined Wireless Networks , 2018, IEEE Wireless Communications.

[12]  K. B. Letaief,et al.  A Survey on Mobile Edge Computing: The Communication Perspective , 2017, IEEE Communications Surveys & Tutorials.

[13]  S. Devipriya,et al.  Improved Max-min heuristic model for task scheduling in cloud , 2013, 2013 International Conference on Green Computing, Communication and Conservation of Energy (ICGCE).

[14]  Nei Kato,et al.  Routing or Computing? The Paradigm Shift Towards Intelligent Computer Network Packet Transmission Based on Deep Learning , 2017, IEEE Transactions on Computers.

[15]  Ji Li,et al.  DSCNN: Hardware-oriented optimization for Stochastic Computing based Deep Convolutional Neural Networks , 2016, 2016 IEEE 34th International Conference on Computer Design (ICCD).

[16]  Yan Wang,et al.  Computation Offloading with Multiple Agents in Edge-Computing–Supported IoT , 2019, ACM Trans. Sens. Networks.

[17]  Eui-nam Huh,et al.  Joint Node Selection and Resource Allocation for Task Offloading in Scalable Vehicle-Assisted Multi-Access Edge Computing , 2019, Symmetry.

[18]  Sheng Chen,et al.  On the Serviceability of Mobile Vehicular Cloudlets in a Large-Scale Urban Environment , 2016, IEEE Transactions on Intelligent Transportation Systems.

[19]  Lihong Li,et al.  PAC model-free reinforcement learning , 2006, ICML.

[20]  Wei Yan,et al.  Optimal Scheduling Algorithm of MapReduce Tasks Based on QoS in the Hybrid Cloud , 2016, 2016 17th International Conference on Parallel and Distributed Computing, Applications and Technologies (PDCAT).

[21]  Yang Yu,et al.  Development of an Efficient Driving Strategy for Connected and Automated Vehicles at Signalized Intersections: A Reinforcement Learning Approach , 2020, IEEE Transactions on Intelligent Transportation Systems.

[22]  Yang Wang,et al.  Adaptive Scheduling of Task Graphs with Dynamic Resilience , 2017, IEEE Transactions on Computers.

[23]  Jin Gon Shon,et al.  An Improvement on the Weighted Least-Connection Scheduling Algorithm for Load Balancing in Web Cluster Systems , 2010, FGIT-GDC/CA.

[24]  Song Guo,et al.  Joint Optimization of Task Scheduling and Image Placement in Fog Computing Supported Software-Defined Embedded System , 2016, IEEE Transactions on Computers.

[25]  Dongrui Fan,et al.  An Evolutionary Technique for Performance-Energy-Temperature Optimized Scheduling of Parallel Tasks on Multi-Core Processors , 2016, IEEE Transactions on Parallel and Distributed Systems.

[26]  Song Guo,et al.  Traffic and Computation Co-Offloading With Reinforcement Learning in Fog Computing for Industrial Applications , 2019, IEEE Transactions on Industrial Informatics.

[27]  Medhat A. Tawfeek,et al.  Cloud task scheduling based on ant colony optimization , 2013, 2013 8th International Conference on Computer Engineering & Systems (ICCES).

[28]  Hang Liu,et al.  Multi-Objective Workflow Scheduling With Deep-Q-Network-Based Multi-Agent Reinforcement Learning , 2019, IEEE Access.

[29]  Lihong Li,et al.  Sample Complexity of Multi-task Reinforcement Learning , 2013, UAI.

[30]  Jiannong Cao,et al.  PVA in VANETs: Stopped cars are not silent , 2011, 2011 Proceedings IEEE INFOCOM.

[31]  Wenzhong Li,et al.  Efficient Multi-User Computation Offloading for Mobile-Edge Cloud Computing , 2015, IEEE/ACM Transactions on Networking.

[32]  Hai Jin,et al.  Developing resource consolidation frameworks for moldable virtual machines in clouds , 2014, Future Gener. Comput. Syst..

[33]  Cheng-Xiang Wang,et al.  Cost Optimization for On-Demand Content Streaming in IoV Networks With Two Service Tiers , 2019, IEEE Internet of Things Journal.

[34]  Traian Rebedea,et al.  Playing Atari Games with Deep Reinforcement Learning and Human Checkpoint Replay , 2016, ArXiv.

[35]  Xin Liu,et al.  Adaptive Learning-Based Task Offloading for Vehicular Edge Computing Systems , 2019, IEEE Transactions on Vehicular Technology.

[36]  Jie Huang,et al.  A Computation Offloading Algorithm Based on Game Theory for Vehicular Edge Networks , 2018, 2018 IEEE International Conference on Communications (ICC).

[37]  Verena Kantere,et al.  Cloud Resource Allocation from the User Perspective: A Bare-Bones Reinforcement Learning Approach , 2016, WISE.

[38]  Meikang Qiu,et al.  Online optimization for scheduling preemptable tasks on IaaS cloud systems , 2012, J. Parallel Distributed Comput..

[39]  Risat Pathan,et al.  Scheduling Parallel Real-Time Recurrent Tasks on Multicore Platforms , 2018, IEEE Transactions on Parallel and Distributed Systems.

[40]  Safia Kedad-Sidhoum,et al.  Scheduling Independent Moldable Tasks on Multi-Cores with GPUs , 2017, IEEE Transactions on Parallel and Distributed Systems.

[41]  Minming Li,et al.  Energy Optimal Task Scheduling with Normally-Off Local Memory and Sleep-Aware Shared Memory with Access Conflict , 2018, IEEE Transactions on Computers.

[42]  Valentin Cristea,et al.  Reputation Guided Genetic Scheduling Algorithm for Independent Tasks in Inter-clouds Environments , 2013, 2013 27th International Conference on Advanced Information Networking and Applications Workshops.

[43]  Lei Guo,et al.  Deep Reinforcement Learning for Intelligent Internet of Vehicles: An Energy-Efficient Computational Offloading Scheme , 2019, IEEE Transactions on Cognitive Communications and Networking.

[44]  Haifeng Lu,et al.  Optimization of lightweight task offloading strategy for mobile edge computing based on deep reinforcement learning , 2020, Future Gener. Comput. Syst..

[45]  Srikanth Kandula,et al.  Resource Management with Deep Reinforcement Learning , 2016, HotNets.

[46]  Peter Stone,et al.  Transfer Learning for Reinforcement Learning Domains: A Survey , 2009, J. Mach. Learn. Res..

[47]  Xiaomin Zhu,et al.  ANGEL: Agent-Based Scheduling for Real-Time Tasks in Virtualized Clouds , 2015, IEEE Transactions on Computers.

[48]  Nan Zhao,et al.  Integrated Networking, Caching, and Computing for Connected Vehicles: A Deep Reinforcement Learning Approach , 2018, IEEE Transactions on Vehicular Technology.