Distributed Independent Reinforcement Learning (DIRL) Approach to Resource Management in Wireless Sensor Networks

In wireless sensor networks, resource-constrained nodes are expected to operate in unattended highly dynamic environments. Hence, the need for adaptive and autonomous resource/task management in wireless sensor networks is well recognized. We present distributed independent reinforcement learning (DIRL), a Q-learning based framework to enable autonomous self-learning/adaptive applications with inherent support for efficient resource/task management. The proposed scheme based on DIRL, learns the utility of performing various tasks over time using mostly local information at nodes and uses the utility value along with application constraints for task management by optimizing global system-wide parameters like total energy usage, network lifetime etc. We also present an object tracking application design based on DIRL to exemplify our framework. Finally, we present results of simulation studies to demonstrate the feasibility of our approach and compare its performance against other existing approaches. In general for applications requiring autonomous adaptation, we show that DIRL on average is about 90% more efficient than traditional resource management schemes like static scheduling without losing any significant accuracy/performance.

[1]  Viktor K. Prasanna,et al.  Issues in designing middleware for wireless sensor networks , 2004, IEEE Network.

[2]  David E. Culler,et al.  Hood: a neighborhood abstraction for sensor networks , 2004, MobiSys '04.

[3]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[4]  Milind Tambe,et al.  Distributed Sensor Networks , 2003, Multiagent Systems, Artificial Societies, and Simulated Organizations.

[5]  Matt Welsh,et al.  Decentralized, adaptive resource allocation for sensor networks , 2005, NSDI.

[6]  Wei Hong,et al.  Proceedings of the 5th Symposium on Operating Systems Design and Implementation Tag: a Tiny Aggregation Service for Ad-hoc Sensor Networks , 2022 .

[7]  Kay Römer,et al.  Algorithms for generic role assignment in wireless sensor networks , 2005, SenSys '05.

[8]  Margaret Martonosi,et al.  Impala: a middleware system for managing autonomic, parallel sensor systems , 2003, PPoPP '03.

[9]  Philip Levis,et al.  Maté: a tiny virtual machine for sensor networks , 2002, ASPLOS X.

[10]  Sandeep Neema,et al.  Constraint-guided dynamic reconfiguration in sensor networks , 2004, Third International Symposium on Information Processing in Sensor Networks, 2004. IPSN 2004.

[11]  Ben J. A. Kröse,et al.  Learning from delayed rewards , 1995, Robotics Auton. Syst..

[12]  Sridhar Mahadevan,et al.  Automatic Programming of Behavior-Based Robots Using Reinforcement Learning , 1991, Artif. Intell..

[13]  Jörg Hähner,et al.  TinyCubus: a flexible and adaptive framework sensor networks , 2005, Proceeedings of the Second European Workshop on Wireless Sensor Networks, 2005..

[14]  Matt Welsh,et al.  Programming Sensor Networks Using Abstract Regions , 2004, NSDI.

[15]  Stephen B. Wicker,et al.  On the Complexity of Distributed Self-Configuration in Wireless Networks , 2003, Telecommun. Syst..

[16]  John W. Byers,et al.  Utility-based decision-making in wireless sensor networks , 2000, 2000 First Annual Workshop on Mobile and Ad Hoc Networking and Computing. MobiHOC (Cat. No.00EX444).

[17]  Amy L. Murphy,et al.  Middleware to support sensor network applications , 2004, IEEE Network.

[18]  Richard S. Sutton,et al.  Temporal credit assignment in reinforcement learning , 1984 .

[19]  Craig Boutilier,et al.  The Dynamics of Reinforcement Learning in Cooperative Multiagent Systems , 1998, AAAI/IAAI.

[20]  Richard R. Brooks,et al.  Distributed Sensor Networks: A Multiagent Perspective , 2008 .