Ember: energy management of batteryless event detection sensors with deep reinforcement learning

Energy management can extend the lifetime of batteryless, energy-harvesting systems by judiciously utilizing the energy available. Duty cycling of such systems is especially challenging for event detection, as events arrive sporadically and energy availability is uncertain. If the node sleeps too much, it may miss important events; if it depletes energy too quickly, it will stop operating in low energy conditions and miss events. Thus, accurate event prediction is important in making this tradeoff. We propose Ember, an energy management system based on deep reinforcement learning to duty cycle event-driven sensors in low energy conditions. We train a policy using historical real-world data traces of motion, temperature, humidity, pressure, and light events. The resulting policy can learn to capture up to 95% of the events without depleting the node. Without historical data for training when deploying a node at a new location, we propose a self-supervised mechanism to collect ground-truth data while learning from the data at the same time. Ember learns to capture the majority of events within a week without any historical data and matches the performance of the policies trained with historical data in a few weeks. We deployed 40 nodes running Ember for indoor sensing and demonstrate that the learned policies generalize to real-world settings as well as outperform state-of-the-art techniques.

[1]  Jacob Sorber,et al.  Tragedy of the Coulombs: Federating Energy Storage for Tiny, Intermittently-Powered Sensors , 2015, SenSys.

[2]  Michael I. Jordan,et al.  RLlib: Abstractions for Distributed Reinforcement Learning , 2017, ICML.

[3]  Rajesh K. Gupta,et al.  Speed control and scheduling of data mules in sensor networks , 2010, TOSN.

[4]  Sergey Levine,et al.  Trust Region Policy Optimization , 2015, ICML.

[5]  Oriol Vinyals,et al.  Representation Learning with Contrastive Predictive Coding , 2018, ArXiv.

[6]  Francesco Fraternali,et al.  Scaling configuration of energy harvesting sensors with reinforcement learning , 2018, ENSsys@SenSys.

[7]  Leon Atkins,et al.  Algorithms for power savings , 2014 .

[8]  Jakub W. Pachocki,et al.  Dota 2 with Large Scale Deep Reinforcement Learning , 2019, ArXiv.

[9]  Peter Dayan,et al.  Q-learning , 1992, Machine Learning.

[10]  Jeremy Gummeson,et al.  SkinnyPower: enabling batteryless wearable sensors via intra-body power transfer , 2019, SenSys.

[11]  Hiroshi Nakamura,et al.  Power Management of Wireless Sensor Nodes with Coordinated Distributed Reinforcement Learning , 2019, 2019 IEEE 37th International Conference on Computer Design (ICCD).

[12]  Prabal Dutta,et al.  Capacity Over Capacitance for Reliable Energy Harvesting Sensors , 2019, 2019 18th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN).

[13]  Brandon Lucia,et al.  Chain: tasks and channels for reliable intermittent programs , 2016, OOPSLA.

[14]  Mani B. Srivastava,et al.  Adaptive Duty Cycling for Energy Harvesting Systems , 2006, ISLPED'06 Proceedings of the 2006 International Symposium on Low Power Electronics and Design.

[15]  Xin Liu,et al.  Passive-ZigBee: Enabling ZigBee Communication in IoT Networks with 1000X+ Less Power Consumption , 2018, SenSys.

[16]  Lihua Tang,et al.  Powering indoor sensing with airflows: a trinity of energy harvesting, synchronous duty-cycling, and sensing , 2013, SenSys '13.

[17]  Nikos Komodakis,et al.  Unsupervised Representation Learning by Predicting Image Rotations , 2018, ICLR.

[18]  Prabal Dutta,et al.  An energy-harvesting sensor architecture and toolkit for building monitoring and event detection , 2014, BuildSys@SenSys.

[19]  Purushottam Kulkarni,et al.  Energy Harvesting Sensor Nodes: Survey and Implications , 2011, IEEE Communications Surveys & Tutorials.

[20]  Brandon Lucia,et al.  Intermittent Computing: Challenges and Opportunities , 2017, SNAPL.

[21]  Eduardo F. Morales,et al.  An Introduction to Reinforcement Learning , 2011 .

[22]  Prabal Dutta,et al.  Cinamin: A Perpetual and Nearly Invisible BLE Beacon , 2016, EWSN.

[23]  Srikanth Kandula,et al.  Resource Management with Deep Reinforcement Learning , 2016, HotNets.

[24]  Luca Benini,et al.  Pible: battery-free mote for perpetual indoor BLE applications , 2018, BuildSys@SenSys.

[25]  Thomas Brox,et al.  Discriminative Unsupervised Feature Learning with Exemplar Convolutional Neural Networks , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[26]  Prabal Dutta,et al.  Monjolo: an energy-harvesting energy meter architecture , 2013, SenSys '13.

[27]  Chuan Zhang,et al.  A Reconfigurable Smart Sensor Interface for Industrial WSN in IoT Environment , 2014, IEEE Transactions on Industrial Informatics.

[28]  Boris Bellalta,et al.  Adapting sampling interval of sensor networks using on-line reinforcement learning , 2016, 2016 IEEE 3rd World Forum on Internet of Things (WF-IoT).

[29]  Kuan-Chieh Wang,et al.  QoS-Aware Power Management for Energy Harvesting Wireless Sensor Network Utilizing Reinforcement Learning , 2009, 2009 International Conference on Computational Science and Engineering.

[30]  Sachin Katti,et al.  HitchHike: Practical Backscatter Using Commodity WiFi , 2016, SenSys.

[31]  Shuguang Cui,et al.  Reinforcement Learning-Based Multiaccess Control and Battery Prediction With Energy Harvesting in IoT Systems , 2018, IEEE Internet of Things Journal.

[32]  Jacob Sorber,et al.  Timely Execution on Intermittently Powered Batteryless Sensors , 2017, SenSys.

[33]  Jeremy Gummeson,et al.  Enabling battery-less wearable sensors via intra-body power transfer: demo abstract , 2019, SenSys.

[34]  Vinod Sharma,et al.  Optimal energy management policies for energy harvesting sensor nodes , 2008, IEEE Transactions on Wireless Communications.

[35]  Alberto Rodriguez,et al.  Learning Synergies Between Pushing and Grasping with Self-Supervised Deep Reinforcement Learning , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[36]  Luca Benini,et al.  Adaptive Power Management for Environmentally Powered Systems , 2010, IEEE Transactions on Computers.

[37]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[38]  Prabal Dutta,et al.  Perpetual Sensing for the Built Environment , 2016, IEEE Pervasive Computing.

[39]  Andrew G. Barto,et al.  Adaptive Control of Duty Cycling in Energy-Harvesting Wireless Sensor Networks , 2007, 2007 4th Annual IEEE Communications Society Conference on Sensor, Mesh and Ad Hoc Communications and Networks.

[40]  Brandon Lucia,et al.  Intelligence Beyond the Edge: Inference on Intermittent Embedded Systems , 2018, ASPLOS.

[41]  Tao Sun,et al.  DeepRacer: Autonomous Racing Platform for Experimentation with Sim2Real Reinforcement Learning , 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[42]  Gaurav S. Sukhatme,et al.  Sim-to-(Multi)-Real: Transfer of Low-Level Robust Control Policies to Multiple Quadrotors , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[43]  Klaus D. McDonald-Maier,et al.  Direct Reinforcement Learning for Autonomous Power Configuration and Control in Wireless Networks , 2009, 2009 NASA/ESA Conference on Adaptive Hardware and Systems.

[44]  Przemyslaw Pawelczak,et al.  InK: Reactive Kernel for Tiny Batteryless Sensors , 2018, SenSys.

[45]  Joshua R. Smith,et al.  Wi-fi backscatter , 2014, SIGCOMM 2015.

[46]  Frank Alexander Kraemer,et al.  Autonomous Management of Energy-Harvesting IoT Nodes Using Deep Reinforcement Learning , 2019, 2019 IEEE 13th International Conference on Self-Adaptive and Self-Organizing Systems (SASO).

[47]  Koen Langendoen,et al.  Continuous Sensing on Intermittent Power , 2020, 2020 19th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN).

[48]  Abhinav Gupta,et al.  Supersizing self-supervision: Learning to grasp from 50K tries and 700 robot hours , 2015, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[49]  Mani B. Srivastava,et al.  Power management in energy harvesting sensor networks , 2007, TECS.

[50]  Alexei A. Efros,et al.  Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[51]  Alec Radford,et al.  Proximal Policy Optimization Algorithms , 2017, ArXiv.

[52]  Olivier Berder,et al.  RLMan: An Energy Manager Based on Reinforcement Learning for Energy Harvesting Wireless Sensor Networks , 2018, IEEE Transactions on Green Communications and Networking.

[53]  Brandon Lucia,et al.  Camaroptera: a Batteryless Long-Range Remote Visual Sensing System , 2019, ENSsys@SenSys.

[54]  Sandy Irani,et al.  Online strategies for dynamic power management in systems with multiple power-saving states , 2003, TECS.

[55]  Matthew S. Reynolds,et al.  BLE-Backscatter: Ultralow-Power IoT Nodes Compatible With Bluetooth 4.0 Low Energy (BLE) Smartphones and Tablets , 2017, IEEE Transactions on Microwave Theory and Techniques.

[56]  Hao-Li Wang,et al.  A Reinforcement Learning-Based ToD Provisioning Dynamic Power Management for Sustainable Operation of Energy Harvesting Wireless Sensor Node , 2014, IEEE Transactions on Emerging Topics in Computing.

[57]  Joshua R. Smith,et al.  WISPCam: A battery-free RFID camera , 2015, 2015 IEEE International Conference on RFID (RFID).

[58]  David Wetherall,et al.  Ambient backscatter: wireless communication out of thin air , 2013, SIGCOMM.

[59]  Jacob Sorber,et al.  Flicker: Rapid Prototyping for the Batteryless Internet-of-Things , 2017, SenSys.

[60]  Roy Chaoming Hsu,et al.  Reinforcement Learning-Based Dynamic Power Management for Energy Harvesting Wireless Sensor Network , 2009, IEA/AIE.

[61]  Thomas Weng,et al.  Occupancy-driven energy management for smart building automation , 2010, BuildSys '10.

[62]  Kok-Lim Alvin Yau,et al.  Reinforcement learning for context awareness and intelligence in wireless networks: Review, new features and open issues , 2012, J. Netw. Comput. Appl..

[63]  Trevor Darrell,et al.  Loss is its own Reward: Self-Supervision for Reinforcement Learning , 2016, ICLR.

[64]  Sergey Levine,et al.  Self-Supervised Deep Reinforcement Learning with Generalized Computation Graphs for Robot Navigation , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[65]  Luca Benini,et al.  Pible: battery-free mote for perpetual indoor BLE applications: demo abstract , 2018, BuildSys@SenSys.

[66]  Sergey Levine,et al.  Diversity is All You Need: Learning Skills without a Reward Function , 2018, ICLR.

[67]  Jakub W. Pachocki,et al.  Learning dexterous in-hand manipulation , 2018, Int. J. Robotics Res..

[68]  Shuguang Cui,et al.  Power Control in Energy Harvesting Multiple Access System With Reinforcement Learning , 2019, IEEE Internet of Things Journal.

[69]  Sergey Levine,et al.  High-Dimensional Continuous Control Using Generalized Advantage Estimation , 2015, ICLR.

[70]  Nitish Srivastava Unsupervised Learning of Visual Representations using Videos , 2015 .

[71]  Donald L. Iglehart,et al.  Importance sampling for stochastic simulations , 1989 .

[72]  Taehoon Kim,et al.  Quantifying Generalization in Reinforcement Learning , 2018, ICML.

[73]  Rajesh K. Gupta,et al.  ACES - Automatic Configuration of Energy Harvesting Sensors with Reinforcement Learning , 2020, ACM Trans. Sens. Networks.