DA-DRLS: Drift adaptive deep reinforcement learning based scheduling for IoT resource management

Abstract In order to fulfill the tremendous resource demand by diverse IoT applications, the large-scale resource-constrained IoT ecosystem requires a robust resource management technique. An optimal resource provisioning in IoT ecosystem deals with an efficient request-resource mapping which is difficult to achieve due to the heterogeneity and dynamicity of IoT resources and IoT requests. In this paper, we investigate the scheduling and resource allocation problem for dynamic user requests with varying resource requirements. Specifically, we formulate the complete problem as an optimization problem and try to generate an optimal policy with the objectives to minimize the overall energy consumption and to achieve a long-term user satisfaction through minimum response time. We introduce the paradigm of a deep reinforcement learning (DRL) mechanism to escalate the resource management efficiency in IoT ecosystem. To maximize the numerical performance of the entire resource management activities, our method learns to select the optimal resource allocation policy among a number of possible solutions. Moreover, the proposed approach can efficiently handle a sudden hike or fall in users' demand, which we call demand drift, through adaptive learning maintaining the optimal resource utilization. Finally, our simulation analysis illustrates the effectiveness of the proposed mechanism as it achieves substantial improvements in various factors, like reducing energy consumption and response time by at least 36.7% and 59.7% respectively and increasing average resource utilization by at least 10.4%. Our approach also attains a good convergence and a trade-off between the monitoring metrics.

[1]  Sherali Zeadally,et al.  Wireless energy harvesting: Empirical results and practical considerations for Internet of Things , 2018, J. Netw. Comput. Appl..

[2]  Razvan Pascanu,et al.  Theano: new features and speed improvements , 2012, ArXiv.

[3]  Zhenchun Wei,et al.  A task scheduling algorithm based on Q-learning and shared value function for WSNs , 2017, Comput. Networks.

[4]  Rashid Mehmood,et al.  Enabling Reliable and Resilient IoT Based Smart City Applications , 2017 .

[5]  Huilong Duan,et al.  Reinforcement learning based resource allocation in business process management , 2011, Data Knowl. Eng..

[6]  Juergen Jasperneite,et al.  The Future of Industrial Communication: Automation Networks in the Era of the Internet of Things and Industry 4.0 , 2017, IEEE Industrial Electronics Magazine.

[7]  B. B. Zaidan,et al.  A review of smart home applications based on Internet of Things , 2017, J. Netw. Comput. Appl..

[8]  Marcello Restelli,et al.  Stochastic Variance-Reduced Policy Gradient , 2018, ICML.

[9]  Valérie Issarny,et al.  Revisiting Service-Oriented Architecture for the IoT: A Middleware Perspective , 2016, ICSOC.

[10]  Wei Xu,et al.  Energy Efficient Resource Allocation in Machine-to-Machine Communications With Multiple Access and Energy Harvesting for IoT , 2017, IEEE Internet of Things Journal.

[11]  K. R. Venugopal,et al.  Searching for the IoT Resources: Fundamentals, Requirements, Comprehensive Review, and Future Directions , 2018, IEEE Communications Surveys & Tutorials.

[12]  Yogesh L. Simmhan,et al.  Distributed Scheduling of Event Analytics across Edge and Cloud , 2016, ACM Trans. Cyber Phys. Syst..

[13]  Mukesh Taneja,et al.  A framework for power saving in IoT networks , 2014, 2014 International Conference on Advances in Computing, Communications and Informatics (ICACCI).

[14]  Ekram Hossain,et al.  Deep Learning for Radio Resource Allocation in Multi-Cell Networks , 2018, IEEE Network.

[15]  Jenq-Shiou Leu,et al.  Improving Heterogeneous SOA-Based IoT Message Stability by Shortest Processing Time Scheduling , 2014, IEEE Transactions on Services Computing.

[16]  Paulo F. Pires,et al.  Resource Management for Internet of Things , 2017, Springer Briefs in Computer Science.

[17]  Dongbin Zhao,et al.  Deep Reinforcement Learning With Visual Attention for Vehicle Classification , 2017, IEEE Transactions on Cognitive and Developmental Systems.

[18]  Chong Shen,et al.  Reinforcement learning models for scheduling in wireless networks , 2013, Frontiers of Computer Science.

[19]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[20]  Vangelis Gazis,et al.  A Survey of Standards for Machine-to-Machine and the Internet of Things , 2017, IEEE Communications Surveys & Tutorials.

[21]  Brigitte Bigi,et al.  Using Kullback-Leibler Distance for Text Categorization , 2003, ECIR.

[22]  Bin Li,et al.  Energy-Efficient User Scheduling and Power Allocation for NOMA-Based Wireless Networks With Massive IoT Devices , 2018, IEEE Internet of Things Journal.

[23]  Shahriar Mirabbasi,et al.  Wireless Energy Harvesting for Internet of Things , 2014 .

[24]  Mohammed Atiquzzaman,et al.  Scheduling internet of things applications in cloud computing , 2016, Annals of Telecommunications.

[25]  Martin Maier,et al.  Power-Saving Methods for Internet of Things over Converged Fiber-Wireless Access Networks , 2016, IEEE Communications Magazine.

[26]  Yishay Mansour,et al.  Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.

[27]  Smruti R. Sarangi,et al.  Energy efficient scheduling in IoT networks , 2018, SAC.

[28]  Bernhard Rinner,et al.  Resource coordination in wireless sensor networks by cooperative reinforcement learning , 2012, 2012 IEEE International Conference on Pervasive Computing and Communications Workshops.

[29]  Jing Wang,et al.  A deep reinforcement learning based framework for power-efficient resource allocation in cloud RANs , 2017, 2017 IEEE International Conference on Communications (ICC).

[30]  Li Li,et al.  Traffic signal timing via deep reinforcement learning , 2016, IEEE/CAA Journal of Automatica Sinica.

[31]  Luciano Bononi,et al.  Adaptive Sensing Scheduling and Spectrum Selection in Cognitive Wireless Mesh Networks , 2011, 2011 Proceedings of 20th International Conference on Computer Communications and Networks (ICCCN).

[32]  P. Venkata Krishna,et al.  Power modelling of sensors for IoT using reinforcement learning , 2018, Int. J. Adv. Intell. Paradigms.

[33]  Rashid Mehmood,et al.  Data Fusion and IoT for Smart Ubiquitous Environments: A Survey , 2017, IEEE Access.

[34]  Srikanth Kandula,et al.  Resource Management with Deep Reinforcement Learning , 2016, HotNets.

[35]  Ling Li,et al.  QoS-Aware Scheduling of Services-Oriented Internet of Things , 2014, IEEE Transactions on Industrial Informatics.

[36]  Navin Kumar,et al.  Packet Scheduling Scheme to Guarantee QoS in Internet of Things , 2018, Wirel. Pers. Commun..

[37]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[38]  Weiwei Lin,et al.  Random task scheduling scheme based on reinforcement learning in cloud computing , 2015, Cluster Computing.

[39]  Sang Won Yoon,et al.  Distributed scheduling using belief propagation for internet-of-things (IoT) networks , 2018, Peer-to-Peer Netw. Appl..

[40]  Marco Pavone,et al.  Cellular Network Traffic Scheduling With Deep Reinforcement Learning , 2018, AAAI.

[41]  Sergey Levine,et al.  High-Dimensional Continuous Control Using Generalized Advantage Estimation , 2015, ICLR.

[42]  Nikos D. Sidiropoulos,et al.  Learning to optimize: Training deep neural networks for wireless resource management , 2017, SPAWC.

[43]  Gang Wang,et al.  Reinforcement Learning for Learning Rate Control , 2017, ArXiv.

[44]  Saima Abdullah,et al.  An Energy Efficient Message Scheduling Algorithm Considering Node Failure in IoT Environment , 2014, Wireless Personal Communications.

[45]  Shamim Nemati,et al.  Optimal medication dosing from suboptimal clinical examples: A deep reinforcement learning approach , 2016, 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC).

[46]  Sergey Levine,et al.  Trust Region Policy Optimization , 2015, ICML.

[47]  Bart De Schutter,et al.  Reinforcement Learning and Dynamic Programming Using Function Approximators , 2010 .

[48]  Mohsen Guizani,et al.  Semisupervised Deep Reinforcement Learning in Support of IoT and Smart City Services , 2018, IEEE Internet of Things Journal.

[49]  Yuan Xue,et al.  Autonomic Joint Session Scheduling Strategies for Heterogeneous Wireless Networks , 2008, 2008 IEEE Wireless Communications and Networking Conference.

[50]  Shai Ben-David,et al.  Detecting Change in Data Streams , 2004, VLDB.

[51]  Abhijeet Bhorkar,et al.  Adaptive Opportunistic Routing for Wireless Ad Hoc Networks , 2012, IEEE/ACM Transactions on Networking.

[52]  Ugur Çetintemel,et al.  Plan-based complex event detection across distributed sources , 2008, Proc. VLDB Endow..

[53]  Marimuthu Palaniswami,et al.  Internet of Things (IoT): A vision, architectural elements, and future directions , 2012, Future Gener. Comput. Syst..

[54]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[55]  Abishi Chowdhury,et al.  A survey study on Internet of Things resource management , 2018, J. Netw. Comput. Appl..

[56]  Debabrata Das,et al.  Efficient Anomaly Detection Methodology for Power Saving in Massive IoT Architecture , 2018, ICDCIT.

[57]  Kuang-Ching Wang,et al.  Review of Internet of Things (IoT) in Electric Power and Energy Systems , 2018, IEEE Internet of Things Journal.

[58]  Xianchun Zhang,et al.  Complex IoT Control System Modeling from Perspectives of Environment Perception and Information Security , 2017, Mobile Networks and Applications.

[59]  Thiemo Voigt,et al.  Velox VM: A safe execution environment for resource-constrained IoT applications , 2018, J. Netw. Comput. Appl..

[60]  Aiiad Albeshri,et al.  Analysis of Eight Data Mining Algorithms for Smarter Internet of Things (IoT) , 2016, EUSPN/ICTH.

[61]  Mohammed Alodib QoS-Aware approach to monitor violations of SLAs in the IoT , 2016, J. Innov. Digit. Ecosyst..

[62]  Song Guo,et al.  Green Industrial Internet of Things Architecture: An Energy-Efficient Perspective , 2016, IEEE Communications Standards.