Optimal Selective Transmission Policy for Energy-Harvesting Wireless Sensors via Monotone Neural Networks

We investigate the optimal transmission policy for an energy-harvesting wireless sensor node. The node must decide whether an arrived packet should be transmitted or dropped, based on the packet’s priority, wireless channel gain, and the energy status of the node. The problem is formulated under the Markov decision process (MDP) framework. For such a problem, the conventional method to get the optimal policy is by using a state value function, which is three-dimensional in the considered problem, leading to high complexity. Fortunately, to reduce complexity, we derive an equivalent solution for the optimal policy via a one-dimensional after-state value function. We show that the after-state value function is differentiable and nondecreasing. We also discover a threshold structure of the optimal policy that is derived by the after-state value function. Furthermore, to approximate the after-state value function, we propose a learning algorithm to train a three-layer monotone neural network. The trained network thus finds a near-optimal selective transmission policy of the node. Finally, through simulation, we demonstrate the learning efficiency of the algorithm and the performance of the learned policy.

[1]  Mihaela van der Schaar,et al.  Structure-Aware Stochastic Control for Transmission Scheduling , 2010, IEEE Transactions on Vehicular Technology.

[2]  Hassaan Khaliq Qureshi,et al.  Energy management in harvesting enabled sensing nodes: Prediction and control , 2019, J. Netw. Comput. Appl..

[3]  Csaba Szepesvári,et al.  Finite-Time Bounds for Fitted Value Iteration , 2008, J. Mach. Learn. Res..

[4]  Purushottam Kulkarni,et al.  Energy Harvesting Sensor Nodes: Survey and Implications , 2011, IEEE Communications Surveys & Tutorials.

[5]  Mohamad Assaad,et al.  Multi -Agent Deep Reinforcement Learning based Power Control for Large Energy Harvesting Networks , 2019, 2019 International Symposium on Modeling and Optimization in Mobile, Ad Hoc, and Wireless Networks (WiOPT).

[6]  Yik-Chung Wu,et al.  Joint CFO and Channel Estimation for OFDM-Based Two-Way Relay Networks , 2011, IEEE Transactions on Wireless Communications.

[7]  Chi Zhou,et al.  Maximum Residual Energy Routing in Wastage-Aware Energy Harvesting Wireless Sensor Networks , 2014, 2014 IEEE 79th Vehicular Technology Conference (VTC Spring).

[8]  Min Dong,et al.  Online Joint Power Control for Two-Hop Wireless Relay Networks With Energy Harvesting , 2018, IEEE Transactions on Signal Processing.

[9]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[10]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[11]  Jeffrey G. Andrews,et al.  The Effect of Fading, Channel Inversion, and Threshold Scheduling on Ad Hoc Networks , 2007, IEEE Transactions on Information Theory.

[12]  Dong In Kim,et al.  Theory and Experiment for Wireless-Powered Sensor Networks: How to Keep Sensors Alive , 2017, IEEE Transactions on Wireless Communications.

[13]  Nei Kato,et al.  Model Predictive Joint Transmit Power Control for Improving System Availability in Energy-Harvesting Wireless Mesh Networks , 2018, IEEE Communications Letters.

[14]  Gang Zhou,et al.  Energy Modeling and Optimization for BSN and WiFi Networks Using Joint Data Rate Adaptation , 2016, Ad Hoc Sens. Wirel. Networks.

[15]  Sherali Zeadally,et al.  Enabling Technologies for Green Internet of Things , 2017, IEEE Systems Journal.

[16]  Neelesh B. Mehta,et al.  Energy-Efficient Detection Using Ordered Transmissions in Energy Harvesting WSNs , 2018, 2018 IEEE International Conference on Communications (ICC).

[17]  T. W. Lambert,et al.  Modern estimation of the parameters of the Weibull wind speed distribution for wind energy analysis , 2000 .

[18]  Duy Trong Ngo,et al.  A Distributed Energy-Harvesting-Aware Routing Algorithm for Heterogeneous IoT Networks , 2018, IEEE Transactions on Green Communications and Networking.

[19]  Mihaela van der Schaar,et al.  Fast Reinforcement Learning for Energy-Efficient Wireless Communication , 2010, IEEE Transactions on Signal Processing.

[20]  Di Xiao,et al.  Energy modeling and optimization through joint packet size analysis of BSN and WiFi networks , 2011, IPCCC.

[21]  Yueping Wu,et al.  Delay-Aware BS Discontinuous Transmission Control and User Scheduling for Energy Harvesting Downlink Coordinated MIMO Systems , 2012, IEEE Transactions on Signal Processing.

[22]  Hassaan Khaliq Qureshi,et al.  Combined Data Rate and Energy Management in Harvesting Enabled Tactile IoT Sensing Devices , 2019, IEEE Transactions on Industrial Informatics.

[23]  Jesús Cid-Sueiro,et al.  An MDP Model for Censoring in Harvesting Sensors: Optimal and Approximated Solutions , 2015, IEEE Journal on Selected Areas in Communications.

[24]  Xiang-Yang Li,et al.  Energy Efficient TDMA Sleep Scheduling in Wireless Sensor Networks , 2009, IEEE INFOCOM 2009.

[25]  Chintha Tellambura,et al.  Channel Estimation and Training Design for Two-Way Relay Networks in Time-Selective Fading Environments , 2011, IEEE Transactions on Wireless Communications.

[26]  Kaibin Huang,et al.  Energy Harvesting Wireless Communications: A Review of Recent Advances , 2015, IEEE Journal on Selected Areas in Communications.

[27]  P. G. Scholar Fast Data Collection in Wireless Sensor Networks , 2015 .

[28]  Michele Zorzi,et al.  On optimal transmission policies for energy harvesting devices , 2012, 2012 Information Theory and Applications Workshop.

[29]  Jesús Cid-Sueiro,et al.  Optimal Selective Transmission under Energy Constraints in Sensor Networks , 2009, IEEE Transactions on Mobile Computing.

[30]  Dayong Ye,et al.  A Self-Adaptive Sleep/Wake-Up Scheduling Approach for Wireless Sensor Networks , 2018, IEEE Transactions on Cybernetics.

[31]  Mérouane Debbah,et al.  Distributed Power Control for Large Energy Harvesting Networks: A Multi-Agent Deep Reinforcement Learning Approach , 2019, IEEE Transactions on Cognitive Communications and Networking.

[32]  Huang Lee,et al.  Wakeup scheduling in wireless sensor networks , 2006, MobiHoc '06.

[33]  Tho Le-Ngoc,et al.  Optimal Stochastic Power Control for Energy Harvesting Systems With Delay Constraints , 2016, IEEE Journal on Selected Areas in Communications.

[34]  David Simplot-Ryl,et al.  Energy-efficient area monitoring for sensor networks , 2004, Computer.

[35]  Longbo Huang Fast-Convergent Learning-Aided Control in Energy Harvesting Networks , 2020, IEEE Transactions on Mobile Computing.

[36]  Gregory E. Bottomley,et al.  Channel estimation in narrowband wireless communication systems , 2001, Wirel. Commun. Mob. Comput..

[37]  Eytan Modiano,et al.  A Calculus Approach to Energy-Efficient Data Transmission With Quality-of-Service Constraints , 2009, IEEE/ACM Transactions on Networking.

[38]  Li Guang-hui Forest Fire Detection System Based on Wireless Sensor Network , 2006 .

[39]  Bhaskar Krishnamachari,et al.  Deep Reinforcement Learning for Dynamic Multichannel Access in Wireless Networks , 2018, IEEE Transactions on Cognitive Communications and Networking.

[40]  R. Srikant,et al.  Asymptotically Optimal Energy-Aware Routing for Multihop Wireless Networks With Renewable Energy Sources , 2007, IEEE/ACM Transactions on Networking.

[41]  Chintha Tellambura,et al.  Joint data detection and channel estimation for OFDM systems , 2006, IEEE Transactions on Communications.

[42]  Hai Jiang,et al.  Sensing, probing, and transmitting strategy for energy harvesting cognitive radio , 2017, 2017 IEEE International Conference on Communications (ICC).

[43]  Shuguang Cui,et al.  Reinforcement Learning-Based Multiaccess Control and Battery Prediction With Energy Harvesting in IoT Systems , 2018, IEEE Internet of Things Journal.

[44]  Martin L. Puterman,et al.  Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .

[45]  Martin A. Riedmiller,et al.  Batch Reinforcement Learning , 2012, Reinforcement Learning.

[46]  Michele Zorzi,et al.  Transmission Policies for Energy Harvesting Sensors with Time-Correlated Energy Supply , 2013, IEEE Transactions on Communications.

[47]  Roy D. Yates,et al.  A generic model for optimizing single-hop transmission policy of replenishable sensors , 2009, IEEE Transactions on Wireless Communications.

[48]  Luis M. Candanedo,et al.  Data driven prediction models of energy use of appliances in a low-energy house , 2017 .

[49]  S. Kannan,et al.  Optimal Selective Forwarding for Energy Saving in Wireless Sensor Networks , 2012 .

[50]  Martin A. Riedmiller Neural Fitted Q Iteration - First Experiences with a Data Efficient Neural Reinforcement Learning Method , 2005, ECML.

[51]  Marina Velikova,et al.  Monotone and Partially Monotone Neural Networks , 2010, IEEE Transactions on Neural Networks.

[52]  Hai Jiang,et al.  Optimal transmission policy in energy harvesting wireless communications: A learning approach , 2017, 2017 IEEE International Conference on Communications (ICC).

[53]  Shuguang Cui,et al.  Energy-Efficient Cooperative Communication Based on Power Control and Selective Single-Relay in Wireless Sensor Networks , 2008, IEEE Transactions on Wireless Communications.

[54]  Mérouane Debbah,et al.  Deep Learning Based Online Power Control for Large Energy Harvesting Networks , 2019, ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[55]  Hai Jiang,et al.  Joint medium access control, routing and energy distribution in multi-hop wireless networks , 2008, IEEE Transactions on Wireless Communications.

[56]  G. Lewicki,et al.  Approximation by Superpositions of a Sigmoidal Function , 2003 .

[57]  H. Kushner,et al.  Stochastic Approximation and Recursive Algorithms and Applications , 2003 .