Heterogeneous Task Offloading and Resource Allocations via Deep Recurrent Reinforcement Learning in Partial Observable Multifog Networks

As wireless services and applications become more sophisticated and require faster and higher capacity networks, there is a need for an efficient management of the execution of increasingly complex tasks based on the requirements of each application. In this regard, fog computing enables the integration of virtualized servers into networks and brings cloud services closer to end devices. In contrast to the cloud server, the computing capacity of fog nodes is limited and thus a single fog node might not be capable of computing-intensive tasks. In this context, task offloading can be particularly useful at the fog nodes by selecting the suitable nodes and proper resource management while guaranteeing the Quality-of-Service (QoS) requirements of the users. This article studies the design of a joint task offloading and resource allocation control for heterogeneous service tasks in multifog nodes systems. This problem is formulated as a partially observable stochastic game, in which each fog node cooperates to maximize the aggregated local rewards while the nodes only have access to local observations. To deal with partial observability, we apply a deep recurrent <inline-formula> <tex-math notation="LaTeX">$Q$ </tex-math></inline-formula>-network (DRQN) approach to approximate the optimal value functions. The solution is then compared to a deep <inline-formula> <tex-math notation="LaTeX">$Q$ </tex-math></inline-formula>-network (DQN) and deep convolutional <inline-formula> <tex-math notation="LaTeX">$Q$ </tex-math></inline-formula>-network (DCQN) approach to evaluate the performance of different neural networks. Moreover, to guarantee the convergence and accuracy of the neural network, an adjusted exploration–exploitation method is adopted. Provided numerical results show that the proposed algorithm can achieve a higher average success rate and lower average overflow than baseline methods.

[1]  Chunxiao Jiang,et al.  Thirty Years of Machine Learning: The Road to Pareto-Optimal Wireless Networks , 2019, IEEE Communications Surveys & Tutorials.

[2]  Sean R Eddy,et al.  What is dynamic programming? , 2004, Nature Biotechnology.

[3]  Mugen Peng,et al.  Deep Reinforcement Learning-Based Mode Selection and Resource Management for Green Fog Radio Access Networks , 2018, IEEE Internet of Things Journal.

[4]  Wojciech Zaremba,et al.  An Empirical Exploration of Recurrent Network Architectures , 2015, ICML.

[5]  Ananth Balashankar,et al.  Software Defined Networking , 2019, 2019 19th International Conference on Sciences and Techniques of Automatic Control and Computer Engineering (STA).

[6]  Eryk Dutkiewicz,et al.  Sustainable Service Allocation Using a Metaheuristic Technique in a Fog Server for Industrial Applications , 2018, IEEE Transactions on Industrial Informatics.

[7]  Mehdi Bennis,et al.  Multi-Tenant Cross-Slice Resource Orchestration: A Deep Reinforcement Learning Approach , 2018, IEEE Journal on Selected Areas in Communications.

[8]  Eryk Dutkiewicz,et al.  Optimal and Fast Real-Time Resource Slicing With Deep Dueling Neural Networks , 2019, IEEE Journal on Selected Areas in Communications.

[9]  Yi Wu,et al.  Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments , 2017, NIPS.

[10]  Marc St-Hilaire,et al.  Model and Algorithms for the Planning of Fog Computing Networks , 2019, IEEE Internet of Things Journal.

[11]  Wenyu Zhang,et al.  Cooperative Fog Computing for Dealing with Big Data in the Internet of Vehicles: Architecture and Hierarchical Resource Management , 2017, IEEE Communications Magazine.

[12]  Hung-Yu Wei,et al.  5G Radio Access Network Design with the Fog Paradigm: Confluence of Communications and Computing , 2017, IEEE Communications Magazine.

[13]  Hamed Haddadi,et al.  Deep Learning in Mobile and Wireless Networking: A Survey , 2018, IEEE Communications Surveys & Tutorials.

[14]  Zhu Han,et al.  Computing Resource Allocation in Three-Tier IoT Fog Networks: A Joint Optimization Approach Combining Stackelberg Game and Matching , 2017, IEEE Internet of Things Journal.

[15]  Lei Li,et al.  Resource Allocation and Task Offloading for Heterogeneous Real-Time Tasks With Uncertain Duration Time in a Fog Queueing System , 2019, IEEE Access.

[16]  Martin L. Puterman,et al.  Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .

[17]  Yi Pan,et al.  Stochastic Load Balancing for Virtual Resource Management in Datacenters , 2020, IEEE Transactions on Cloud Computing.

[18]  Tao Zhang,et al.  Fog and IoT: An Overview of Research Opportunities , 2016, IEEE Internet of Things Journal.

[19]  Roch H. Glitho,et al.  A Comprehensive Survey on Fog Computing: State-of-the-Art and Research Challenges , 2017, IEEE Communications Surveys & Tutorials.

[20]  Bart De Schutter,et al.  A Comprehensive Survey of Multiagent Reinforcement Learning , 2008, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[21]  Raj Jain,et al.  Network virtualization and software defined networking for cloud computing: a survey , 2013, IEEE Communications Magazine.

[22]  Arun Kumar Yadav,et al.  An architecture for Load Balancing Techniques for Fog Computing Environment , 2015 .

[23]  Jiangchuan Liu,et al.  Statistics and Social Network of YouTube Videos , 2008, 2008 16th Interntional Workshop on Quality of Service.

[24]  Joel J. P. C. Rodrigues,et al.  Metaheuristic Scheduling for Cloud: A Survey , 2014, IEEE Systems Journal.

[25]  Rajkumar Buyya,et al.  Internet of Things: Principles and Paradigms , 2016 .

[26]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[27]  Deze Zeng,et al.  Dependency-Aware Computation Offloading in Mobile Edge Computing: A Reinforcement Learning Approach , 2019, IEEE Access.

[28]  Walid Saad,et al.  An Online Optimization Framework for Distributed Fog Network Formation With Minimal Latency , 2017, IEEE Transactions on Wireless Communications.

[29]  Daniel Kudenko,et al.  Reinforcement learning of coordination in cooperative multi-agent systems , 2002, AAAI/IAAI.

[30]  Xingming Sun,et al.  Dynamic Resource Allocation for Load Balancing in Fog Environment , 2018, Wirel. Commun. Mob. Comput..

[31]  Craig Boutilier,et al.  The Dynamics of Reinforcement Learning in Cooperative Multiagent Systems , 1998, AAAI/IAAI.

[32]  Riti Gour,et al.  On Reducing IoT Service Delay via Fog Offloading , 2018, IEEE Internet of Things Journal.

[33]  Mehdi Bennis,et al.  Performance Optimization in Mobile-Edge Computing via Deep Reinforcement Learning , 2018, 2018 IEEE 88th Vehicular Technology Conference (VTC-Fall).

[34]  Sherali Zeadally,et al.  Offloading in fog computing for IoT: Review, enabling technologies, and research opportunities , 2018, Future Gener. Comput. Syst..

[35]  Zdenek Becvar,et al.  Mobile Edge Computing: A Survey on Architecture and Computation Offloading , 2017, IEEE Communications Surveys & Tutorials.

[36]  Olav N. Østerbø,et al.  Modelling of OpenFlow-based software-defined networks: the multiple node case , 2015, IET Networks.

[37]  Wei Yu,et al.  An introduction to convex optimization for communications and signal processing , 2006, IEEE Journal on Selected Areas in Communications.

[38]  Georgios B. Giannakis,et al.  Deep Reinforcement Learning for Adaptive Caching in Hierarchical Content Delivery Networks , 2019, IEEE Transactions on Cognitive Communications and Networking.

[39]  Mugen Peng,et al.  Application of Machine Learning in Wireless Networks: Key Techniques and Open Issues , 2018, IEEE Communications Surveys & Tutorials.

[40]  Mohsen Guizani,et al.  Internet of Things: A Survey on Enabling Technologies, Protocols, and Applications , 2015, IEEE Communications Surveys & Tutorials.

[41]  Mugen Peng,et al.  A Game-Theoretic Approach to Cache and Radio Resource Management in Fog Radio Access Networks , 2019, IEEE Transactions on Vehicular Technology.

[42]  Keith Kirkpatrick,et al.  Software-defined networking , 2013, CACM.

[43]  Xiangjian He,et al.  Using swarm intelligence to optimize the energy consumption for distributed systems , 2013 .

[44]  Victor C. M. Leung,et al.  Deep-Reinforcement-Learning-Based Optimization for Cache-Enabled Opportunistic Interference Alignment Wireless Networks , 2017, IEEE Transactions on Vehicular Technology.

[45]  Wu Muqing,et al.  Multi-Controller Deployment Algorithm in Hierarchical Architecture for SDWAN , 2019, IEEE Access.

[46]  Prasant Mohapatra,et al.  Processor-Network Speed Scaling for Energy–Delay Tradeoff in Smartphone Applications , 2016, IEEE/ACM Transactions on Networking.

[47]  Georges Kaddoum,et al.  Managing Fog Networks using Reinforcement Learning Based Load Balancing Algorithm , 2019, 2019 IEEE Wireless Communications and Networking Conference (WCNC).

[48]  Mugen Peng,et al.  Fog-computing-based radio access networks: issues and challenges , 2015, IEEE Network.

[49]  Chadi Assi,et al.  Dynamic Task Offloading and Scheduling for Low-Latency IoT Services in Multi-Access Edge Computing , 2019, IEEE Journal on Selected Areas in Communications.

[50]  Qianbin Chen,et al.  Computation Offloading and Resource Allocation in Wireless Cellular Networks With Mobile Edge Computing , 2017, IEEE Transactions on Wireless Communications.

[51]  Min Luo,et al.  A Framework for QoS-aware Traffic Classification Using Semi-supervised Machine Learning in SDNs , 2016, 2016 IEEE International Conference on Services Computing (SCC).

[52]  Eduardo F. Morales,et al.  An Introduction to Reinforcement Learning , 2011 .

[53]  Ying-Chang Liang,et al.  Applications of Deep Reinforcement Learning in Communications and Networking: A Survey , 2018, IEEE Communications Surveys & Tutorials.

[54]  Hyundong Shin,et al.  Power Allocation in Cache-Aided NOMA Systems: Optimization and Deep Reinforcement Learning Approaches , 2019, IEEE Transactions on Communications.

[55]  Igor Radusinovic,et al.  Software-Defined Fog Network Architecture for IoT , 2016, Wireless Personal Communications.

[56]  Jonathan P. How,et al.  Deep Decentralized Multi-task Multi-Agent Reinforcement Learning under Partial Observability , 2017, ICML.

[57]  Yoshua Bengio,et al.  Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.

[58]  Peter Stone,et al.  Deep Recurrent Q-Learning for Partially Observable MDPs , 2015, AAAI Fall Symposia.