Sensor Scheduling for Energy-Efficient Target Tracking in Sensor Networks

In this paper, we study the problem of tracking an object moving randomly through a network of wireless sensors. Our objective is to devise strategies for scheduling the sensors to optimize the tradeoff between tracking performance and energy consumption. We cast the scheduling problem as a partially observable Markov decision process (POMDP), where the control actions correspond to the set of sensors to activate at each time step. Using a bottom-up approach, we consider different sensing, motion and cost models with increasing levels of difficulty. At the first level, the sensing regions of the different sensors do not overlap and the target is only observed within the sensing range of an active sensor. Then, we consider sensors with overlapping sensing range such that the tracking error, and hence the actions of the different sensors, are tightly coupled. Finally, we consider scenarios wherein the target locations and sensors' observations assume values on continuous spaces. Exact solutions are generally intractable even for the simplest models due to the dimensionality of the information and action spaces. Hence, we devise approximate solution techniques, and in some cases derive lower bounds on the optimal tradeoff curves. The generated scheduling policies, albeit suboptimal, often provide close-to-optimal energy-tracking tradeoffs.

[1]  G. Monahan State of the Art—A Survey of Partially Observable Markov Decision Processes: Theory, Models, and Algorithms , 1982 .

[2]  John W. Fisher,et al.  Approximate Dynamic Programming for Communication-Constrained Sensor Network Management , 2007, IEEE Transactions on Signal Processing.

[3]  Lihua Xie,et al.  Multi-Sensor Scheduling for Reliable Target Tracking in Wireless Sensor Networks , 2006, 2006 6th International Conference on ITS Telecommunications.

[4]  Leslie Pack Kaelbling,et al.  Learning Policies for Partially Observable Environments: Scaling Up , 1997, ICML.

[5]  Leslie Pack Kaelbling,et al.  Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..

[6]  Edwin K. P. Chong,et al.  Sensor scheduling for target tracking: A Monte Carlo sampling approach , 2006, Digit. Signal Process..

[7]  Darryl Morrell,et al.  Nonmyopic Sensor Scheduling and its Efficient Implementation for Target Tracking Applications , 2006, EURASIP J. Adv. Signal Process..

[8]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.

[9]  Alfred O. Hero,et al.  Sensor management using an active sensing approach , 2005, Signal Process..

[10]  H. Vincent Poor,et al.  An Introduction to Signal Detection and Estimation , 1994, Springer Texts in Electrical Engineering.

[11]  H. Vincent Poor,et al.  An introduction to signal detection and estimation (2nd ed.) , 1994 .

[12]  Jesse Hoey,et al.  Solving POMDPs with Continuous or Large Discrete Observation Spaces , 2005, IJCAI.

[13]  D. Castañón Approximate dynamic programming for sensor management , 1997, Proceedings of the 36th IEEE Conference on Decision and Control.

[14]  Venugopal V. Veeravalli,et al.  Smart Sleeping Policies for Energy Efficient Tracking in Sensor Networks , 2008, IEEE Transactions on Signal Processing.

[15]  Michael L. Littman,et al.  Incremental Pruning: A Simple, Fast, Exact Method for Partially Observable Markov Decision Processes , 1997, UAI.

[16]  Joelle Pineau,et al.  Point-based value iteration: An anytime algorithm for POMDPs , 2003, IJCAI.

[17]  Alfred O. Hero,et al.  A Bayesian method for integrated multitarget tracking and sensor management , 2003, Sixth International Conference of Information Fusion, 2003. Proceedings of the.

[18]  Nicholas Roy,et al.  Exponential Family PCA for Belief Compression in POMDPs , 2002, NIPS.

[19]  Milos Hauskrecht,et al.  Value-Function Approximations for Partially Observable Markov Decision Processes , 2000, J. Artif. Intell. Res..

[20]  ZhaoFeng,et al.  Collaborative in-network processing for target tracking , 2003 .

[21]  Feng Zhao,et al.  Collaborative In-Network Processing for Target Tracking , 2003, EURASIP J. Adv. Signal Process..

[22]  George E. Monahan,et al.  A Survey of Partially Observable Markov Decision Processes: Theory, Models, and Algorithms , 2007 .

[23]  M. Kalandros Covariance control for multisensor systems , 2002 .

[24]  Feng Zhao,et al.  Information-driven dynamic sensor collaboration , 2002, IEEE Signal Process. Mag..

[25]  Nikos A. Vlassis,et al.  Perseus: Randomized Point-based Value Iteration for POMDPs , 2005, J. Artif. Intell. Res..

[26]  Alf Isaksson,et al.  On sensor scheduling via information theoretic criteria , 1999, Proceedings of the 1999 American Control Conference (Cat. No. 99CH36251).

[27]  Dimitri P. Bertsekas,et al.  Dynamic Programming and Optimal Control, Two Volume Set , 1995 .

[28]  Bernard C. Levy,et al.  Principles of Signal Detection and Parameter Estimation , 2008 .

[29]  Edward J. Sondik,et al.  The optimal control of par-tially observable Markov processes , 1971 .

[30]  Andrew W. Moore,et al.  Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..

[31]  Pascal Poupart,et al.  Point-Based Value Iteration for Continuous POMDPs , 2006, J. Mach. Learn. Res..

[32]  N. Zhang,et al.  Algorithms for partially observable markov decision processes , 2001 .