Dynamic power management strategies for a sensor node optimised by reinforcement learning

Dynamic power management DPM is considered in wireless sensor networks for improving the energy efficiency of sensor nodes. DPM usually includes two classical problems, dynamic operating mode OM management and adaptive transmission mechanism. In this paper, we propose a new model of Markov decision process that combines dynamic OM management and adaptive transmission mechanism. In addition, a fragment transmission scheme is further integrated to reduce the probability of retransmission failure and improve the data transmission rate. The model takes into account the performance criteria being the expected cost of synthesising the per-packet energy consumption, the buffer overflow, the fragment cost and the energy consumption of operating mode switching. Reinforcement learning algorithm is subsequently proposed to search for the optimal strategies. Furthermore, a state-clustering approach is given to increase the learning speed and lessen the storage requirements. Finally, an example is presented to illustrate the effectiveness of the proposed method and show that the energy consumption is well-balanced dynamically under the optimised policy, while the throughput decreases only slightly. Therefore, the lifetime of the node can be extended under constrained resources.

[1]  Andrea Bonarini,et al.  A bird's eye view on reinforcement learning approaches for power management in WSNs , 2013, 6th Joint IFIP Wireless and Mobile Networking Conference (WMNC).

[2]  S. Srinivasan,et al.  RFID sensor network-based automation system for monitoring and tracking of sandalwood trees , 2013, Int. J. Comput. Sci. Eng..

[3]  Cory C. Beard,et al.  Analytical models for understanding space, backoff, and flow correlation in CSMA wireless networks , 2012, Wireless Networks.

[4]  Fang Wang,et al.  Cooperative Transmission with Broadcasting and Communication , 2012, 2012 Fourth International Conference on Intelligent Networking and Collaborative Systems.

[5]  Yilun Shang,et al.  Fast distributed consensus seeking in large-scale sensor networks via shortcuts , 2012, Int. J. Comput. Sci. Eng..

[6]  W. Dargie,et al.  Dynamic Power Management in Wireless Sensor Networks: State-of-the-Art , 2012, IEEE Sensors Journal.

[7]  Bernhard Rinner,et al.  Resource coordination in wireless sensor networks by cooperative reinforcement learning , 2012, 2012 IEEE International Conference on Pervasive Computing and Communications Workshops.

[8]  Cory C. Beard,et al.  Competition, cooperation, and optimization in Multi-Hop CSMA networks , 2011, PE-WASUN '11.

[9]  Roy Chaoming Hsu,et al.  Dynamic power management utilizing reinforcement learning with fuzzy reward for energy harvesting wireless sensor nodes , 2011, IECON 2011 - 37th Annual Conference of the IEEE Industrial Electronics Society.

[10]  Somayeh Kianpisheh,et al.  A new approach for power management in sensor node based on reinforcement learning , 2011, 2011 International Symposium on Computer Networks and Distributed Systems (CNDS).

[11]  Somayeh Kianpisheh,et al.  A Power Control Mechanism for Sensor Node Based on Dynamic Programming , 2010, 2010 Second International Conference on Communication Software and Networks.

[12]  Tang Hao,et al.  Look-ahead control of conveyor-serviced production station by using potential-based online policy iteration , 2009, Int. J. Control.

[13]  Prasun Sinha,et al.  Pushback: A hidden Markov model based scheme for energy efficient data transmission in sensor networks , 2009, Ad Hoc Networks.

[14]  Naixue Xiong,et al.  Dynamic power management in new architecture of wireless sensor networks , 2009, Int. J. Commun. Syst..

[15]  Mehul Motani,et al.  Cross-layer Adaptive Transmission: Optimal Strategies in Fading Channels , 2008, IEEE Transactions on Communications.

[16]  Jeong Geun Kim,et al.  Performance Analysis of an Opportunistic Transmission Scheme for Wireless Sensor Networks , 2008, 2008 IEEE International Conference on Communications.

[17]  Xi-Ren Cao,et al.  Stochastic learning and optimization - A sensitivity-based approach , 2007, Annu. Rev. Control..

[18]  Jeong Geun Kim,et al.  Opportunistic Transmission for Wireless Sensor Networks Under Delay Constraints , 2007, ICCSA.

[19]  Jeong Geun Kim,et al.  An Energy-Efficient Transmission Strategy for Wireless Sensor Networks , 2007, 2007 IEEE Wireless Communications and Networking Conference.

[20]  Yang Xiao,et al.  A Survey of Energy-Efficient Scheduling Mechanisms in Sensor Networks , 2006, Mob. Networks Appl..

[21]  K. J. Ray Liu,et al.  Near-optimal reinforcement learning framework for energy-aware sensor communications , 2005, IEEE Journal on Selected Areas in Communications.

[22]  Deborah Estrin,et al.  Medium access control with coordinated adaptive sleeping for wireless sensor networks , 2004, IEEE/ACM Transactions on Networking.

[23]  Ramesh R. Rao,et al.  Improving energy saving in wireless systems by using dynamic power management , 2003, IEEE Trans. Wirel. Commun..

[24]  Anantha Chandrakasan,et al.  Dynamic Power Management in Wireless Sensor Networks , 2001, IEEE Des. Test Comput..

[25]  Saleem A. Kassam,et al.  Finite-state Markov model for Rayleigh fading channels , 1999, IEEE Trans. Commun..

[26]  Dimitri P. Bertsekas,et al.  Dynamic Programming and Optimal Control, Two Volume Set , 1995 .

[27]  Hong Shen Wang,et al.  Finite-state Markov channel-a useful model for radio communication channels , 1995 .