Dynamic power management strategies for a sensor node optimised by reinforcement learning

Dynamic power management DPM is considered in wireless sensor networks for improving the energy efficiency of sensor nodes. DPM usually includes two classical problems, dynamic operating mode OM management and adaptive transmission mechanism. In this paper, we propose a new model of Markov decision process that combines dynamic OM management and adaptive transmission mechanism. In addition, a fragment transmission scheme is further integrated to reduce the probability of retransmission failure and improve the data transmission rate. The model takes into account the performance criteria being the expected cost of synthesising the per-packet energy consumption, the buffer overflow, the fragment cost and the energy consumption of operating mode switching. Reinforcement learning algorithm is subsequently proposed to search for the optimal strategies. Furthermore, a state-clustering approach is given to increase the learning speed and lessen the storage requirements. Finally, an example is presented to illustrate the effectiveness of the proposed method and show that the energy consumption is well-balanced dynamically under the optimised policy, while the throughput decreases only slightly. Therefore, the lifetime of the node can be extended under constrained resources.

[1]  Cory C. Beard,et al.  Analytical models for understanding space, backoff, and flow correlation in CSMA wireless networks , 2012, Wireless Networks.

[2]  Dimitri P. Bertsekas,et al.  Dynamic Programming and Optimal Control, Two Volume Set , 1995 .

[3]  Andrea Bonarini,et al.  A bird's eye view on reinforcement learning approaches for power management in WSNs , 2013, 6th Joint IFIP Wireless and Mobile Networking Conference (WMNC).

[4]  Roy Chaoming Hsu,et al.  Dynamic power management utilizing reinforcement learning with fuzzy reward for energy harvesting wireless sensor nodes , 2011, IECON 2011 - 37th Annual Conference of the IEEE Industrial Electronics Society.

[5]  Tang Hao,et al.  Look-ahead control of conveyor-serviced production station by using potential-based online policy iteration , 2009, Int. J. Control.

[6]  Jeong Geun Kim,et al.  Opportunistic Transmission for Wireless Sensor Networks Under Delay Constraints , 2007, ICCSA.

[7]  Yilun Shang,et al.  Fast distributed consensus seeking in large-scale sensor networks via shortcuts , 2012, Int. J. Comput. Sci. Eng..

[8]  Naixue Xiong,et al.  Dynamic power management in new architecture of wireless sensor networks , 2009 .

[9]  Saleem A. Kassam,et al.  Finite-state Markov model for Rayleigh fading channels , 1999, IEEE Trans. Commun..

[10]  Ramesh R. Rao,et al.  Improving energy saving in wireless systems by using dynamic power management , 2003, IEEE Trans. Wirel. Commun..

[11]  Jeong Geun Kim,et al.  An Energy-Efficient Transmission Strategy for Wireless Sensor Networks , 2007, 2007 IEEE Wireless Communications and Networking Conference.

[12]  W. Dargie,et al.  Dynamic Power Management in Wireless Sensor Networks: State-of-the-Art , 2012, IEEE Sensors Journal.

[13]  Deborah Estrin,et al.  Medium access control with coordinated adaptive sleeping for wireless sensor networks , 2004, IEEE/ACM Transactions on Networking.

[14]  Somayeh Kianpisheh,et al.  A new approach for power management in sensor node based on reinforcement learning , 2011, 2011 International Symposium on Computer Networks and Distributed Systems (CNDS).

[15]  Prasun Sinha,et al.  Pushback: A hidden Markov model based scheme for energy efficient data transmission in sensor networks , 2009, Ad Hoc Networks.

[16]  Xi-Ren Cao,et al.  Stochastic Learning and Optimization: A Sensitivity-Based Approach (International Series on Discrete Event Dynamic Systems) , 2007 .

[17]  Hong Shen Wang,et al.  Finite-state Markov channel-a useful model for radio communication channels , 1995 .

[18]  Fang Wang,et al.  Cooperative Transmission with Broadcasting and Communication , 2012, 2012 Fourth International Conference on Intelligent Networking and Collaborative Systems.

[19]  Anantha Chandrakasan,et al.  Dynamic Power Management in Wireless Sensor Networks , 2001, IEEE Des. Test Comput..

[20]  Mehul Motani,et al.  Cross-layer Adaptive Transmission: Optimal Strategies in Fading Channels , 2008, IEEE Transactions on Communications.

[21]  Bernhard Rinner,et al.  Resource coordination in wireless sensor networks by cooperative reinforcement learning , 2012, 2012 IEEE International Conference on Pervasive Computing and Communications Workshops.

[22]  K. J. Ray Liu,et al.  Near-optimal reinforcement learning framework for energy-aware sensor communications , 2005, IEEE Journal on Selected Areas in Communications.

[23]  Cory C. Beard,et al.  Competition, cooperation, and optimization in Multi-Hop CSMA networks , 2011, PE-WASUN '11.

[24]  Jeong Geun Kim,et al.  Performance Analysis of an Opportunistic Transmission Scheme for Wireless Sensor Networks , 2008, 2008 IEEE International Conference on Communications.

[25]  S. Srinivasan,et al.  RFID sensor network-based automation system for monitoring and tracking of sandalwood trees , 2013, Int. J. Comput. Sci. Eng..

[26]  Yang Xiao,et al.  A Survey of Energy-Efficient Scheduling Mechanisms in Sensor Networks , 2006, Mob. Networks Appl..

[27]  Somayeh Kianpisheh,et al.  A Power Control Mechanism for Sensor Node Based on Dynamic Programming , 2010, 2010 Second International Conference on Communication Software and Networks.