Reinforcement Learning for Self-organizing Wake-Up Scheduling in Wireless Sensor Networks

Wake-up scheduling is a challenging problem in wireless sensor networks. It was recently shown that a promising approach for solving this problem is to rely on reinforcement learning (RL). The RL approach is particularly attractive since it allows the sensor nodes to coordinate through local interactions alone, without the need of central mediator or any form of explicit coordination. This article extends previous work by experimentally studying the behavior of RL wake-up scheduling on a set of three different network topologies, namely line, mesh and grid topologies. The experiments are run using OMNET++, a the state-of-the-art network simulator. The obtained results show how simple and computationally bounded sensor nodes are able to coordinate their wake-up cycles in a distributed way in order to improve the global system performance. The main insight of these experiments is to show that sensor nodes learn to synchronize if they have to cooperate for forwarding data, and learn to desynchronize in order to avoid interferences. This synchronization/desynchronization behavior, referred to for short as (de)synchronicity, allows to improve the message throughput even for very low duty cycles.

[1]  Ivan Stojmenović,et al.  Handbook of Sensor Networks: Algorithms and Architectures , 2005, Handbook of Sensor Networks.

[2]  Shan Liang,et al.  Passive Wake-up Scheme for Wireless Sensor Networks , 2007, Second International Conference on Innovative Computing, Informatio and Control (ICICIC 2007).

[3]  Prasant Mohapatra,et al.  Medium access control in wireless sensor networks , 2007, Comput. Networks.

[4]  Zhenzhen Liu,et al.  RL-MAC: a reinforcement learning based MAC protocol for wireless sensor networks , 2006, Int. J. Sens. Networks.

[5]  Deborah Estrin,et al.  Medium access control with coordinated adaptive sleeping for wireless sensor networks , 2004, IEEE/ACM Transactions on Networking.

[6]  M. Posch,et al.  Win-stay, lose-shift strategies for repeated games-memory length, aspiration levels and noise. , 1999, Journal of theoretical biology.

[7]  Curt Schurgers,et al.  Wakeup Strategies in Wireless Sensor Networks , 2008 .

[8]  Reuven Cohen,et al.  An Optimal Wake-Up Scheduling Algorithm for Minimizing Energy Consumption While Limiting Maximum Delay in a Mesh Sensor Network , 2009, IEEE/ACM Transactions on Networking.

[9]  S. Sitharama Iyengar,et al.  Random asynchronous wakeup protocol for sensor networks , 2004, First International Conference on Broadband Networks.

[10]  J.A. Gutierrez,et al.  IEEE 802.15.4: a developing standard for low-power low-cost wireless personal area networks , 2001, IEEE Network.

[11]  Ann Nowé,et al.  Distributed cooperation in wireless sensor networks , 2011, AAMAS.

[12]  Weili Wu,et al.  Wireless Sensor Networks and Applications , 2008 .

[13]  Chris Watkins,et al.  Learning from delayed rewards , 1989 .

[14]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[15]  Yu-Chee Tseng,et al.  Positioning and location tracking in wireless sensor networks , 2004 .

[16]  Ann Nowé,et al.  Self-organizing Synchronicity and Desynchronicity using Reinforcement Learning , 2011, ICAART.