Collaborative duty cycling strategies in energy harvesting sensor networks

Energy harvesting wireless sensor networks are a promising solution for low cost, long lasting civil monitoring applications. But management of energy consumption is a critical concern to ensure these systems provide maximal utility. Many common civil applications of these networks are fundamentally concerned with detecting and analyzing infrequently occurring events. To conserve energy in these situations, a subset of nodes in the network can assume active duty, listening for events of interest, while the remaining nodes enter low power sleep mode to conserve battery. However, judicious planning of the sequence of active node assignments is needed to ensure that as many nodes as possible can be reached upon the detection of an event, and that the system maintains capability in times of low energy harvesting capabilities. In this article, we propose a novel reinforcement learning (RL) agent, which acts as a centralized power manager for this system. We develop a comprehensive simulation environment to emulate the behavior of an energy harvesting sensor network, with consideration of spatially varying energy harvesting capabilities, and wireless connectivity. We then train the proposed RL agent to learn optimal node selection strategies through interaction with the simulation environment. The behavior and performance of these strategies are tested on real unseen solar energy data, to demonstrate the efficacy of the method. The deep RL agent is shown to outperform baseline approaches on both seen and unseen data.

[1]  Tommaso Addabbo,et al.  A city-scale IoT architecture for monumental structures monitoring , 2019, Measurement.

[2]  Yong Huang,et al.  Multitask Sparse Bayesian Learning with Applications in Structural Health Monitoring , 2018, Comput. Aided Civ. Infrastructure Eng..

[3]  François Chollet,et al.  Keras: The Python Deep Learning library , 2018 .

[4]  Jun Teng,et al.  Cluster-based optimal wireless sensor deployment for structural health monitoring , 2018 .

[5]  Hojjat Adeli,et al.  A novel unsupervised deep learning model for global and local health condition assessment of structures , 2018 .

[6]  Tom Schaul,et al.  Deep Q-learning From Demonstrations , 2017, AAAI.

[7]  Hiroshi Nakamura,et al.  Adaptive Power Management in Solar Energy Harvesting Sensor Node Using Reinforcement Learning , 2017, ACM Trans. Embed. Comput. Syst..

[8]  Hojjat Adeli,et al.  A novel machine learning‐based algorithm to detect damage in high‐rise building structures , 2017 .

[9]  Emad Alsusa,et al.  DC-LEACH: A duty-cycle based clustering protocol for energy harvesting WSNs , 2017, 2017 13th International Wireless Communications and Mobile Computing Conference (IWCMC).

[10]  Aron Habte,et al.  Evaluation of the National Solar Radiation Database (NSRDB): 1998-2015 , 2017 .

[11]  Wojciech Zaremba,et al.  OpenAI Gym , 2016, ArXiv.

[12]  Shaojie Tang,et al.  Enabling Reliable and Network-Wide Wakeup in Wireless Sensor Networks , 2016, IEEE Transactions on Wireless Communications.

[13]  Alex Graves,et al.  Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.

[14]  David Silver,et al.  Deep Reinforcement Learning with Double Q-Learning , 2015, AAAI.

[15]  M. Penrose CONNECTIVITY OF SOFT RANDOM GEOMETRIC GRAPHS , 2013, 1311.3897.

[16]  Pengfei Zhang,et al.  Adaptive Duty Cycling in Sensor Networks With Energy Harvesting Using Continuous-Time Markov Chain and Fluid Models , 2015, IEEE Journal on Selected Areas in Communications.

[17]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[18]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[19]  Hao-Li Wang,et al.  A Reinforcement Learning-Based ToD Provisioning Dynamic Power Management for Sustainable Operation of Energy Harvesting Wireless Sensor Node , 2014, IEEE Transactions on Emerging Topics in Computing.

[20]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[21]  Bhaskar Krishnamachari,et al.  Energy‐efficient deployment strategies in structural health monitoring using wireless sensor networks , 2013 .

[22]  Yuhan Dong,et al.  An effective routing protocol for energy harvesting wireless sensor networks , 2013, 2013 IEEE Wireless Communications and Networking Conference (WCNC).

[23]  Aron Dobos,et al.  System Advisor Model, SAM 2011.12.2: General Description , 2012 .

[24]  Sung-Han Sim,et al.  Decentralized random decrement technique for efficient data aggregation and system identification in wireless smart sensor networks , 2011 .

[25]  Gul Agha,et al.  Flexible smart sensor framework for autonomous structural health monitoring , 2010 .

[26]  Gul Agha,et al.  Structural health monitoring of a cable-stayed bridge using smart sensor technology: deployment and evaluation , 2010 .

[27]  Peter Stone,et al.  Transfer Learning for Reinforcement Learning Domains: A Survey , 2009, J. Mach. Learn. Res..

[28]  Stefan Schaal,et al.  2008 Special Issue: Reinforcement learning of motor skills with policy gradients , 2008 .

[29]  KasabovNikola,et al.  2008 Special issue , 2008 .

[30]  Andrew G. Barto,et al.  Adaptive Control of Duty Cycling in Energy-Harvesting Wireless Sensor Networks , 2007, 2007 4th Annual IEEE Communications Society Conference on Sensor, Mesh and Ad Hoc Communications and Networks.

[31]  Mani B. Srivastava,et al.  Harvesting aware power management for sensor networks , 2006, 2006 43rd ACM/IEEE Design Automation Conference.

[32]  Billie F. Spencer,et al.  Distributed computing strategy for structural health monitoring , 2006 .

[33]  Gary C. Hart,et al.  The structural design of tall and special buildings , 2005 .

[34]  Rong-Hong Jan,et al.  Energy-aware, load balanced routing schemes for sensor networks , 2004, Proceedings. Tenth International Conference on Parallel and Distributed Systems, 2004. ICPADS 2004..

[35]  Anantha P. Chandrakasan,et al.  An application-specific protocol architecture for wireless microsensor networks , 2002, IEEE Trans. Wirel. Commun..

[36]  Jan M. Rabaey,et al.  Energy aware routing for low energy ad hoc sensor networks , 2002, 2002 IEEE Wireless Communications and Networking Conference Record. WCNC 2002 (Cat. No.02TH8609).

[37]  Christos Faloutsos,et al.  Analysis of the Clustering Properties of the Hilbert Space-Filling Curve , 2001, IEEE Trans. Knowl. Data Eng..

[38]  C. R. Dietrich,et al.  Fast and Exact Simulation of Stationary Gaussian Processes through Circulant Embedding of the Covariance Matrix , 1997, SIAM J. Sci. Comput..

[39]  Ben J. A. Kröse,et al.  Learning from delayed rewards , 1995, Robotics Auton. Syst..

[40]  BERNARD M. WAXMAN,et al.  Routing of multipoint connections , 1988, IEEE J. Sel. Areas Commun..

[41]  R. Bellman A Markovian Decision Process , 1957 .