Sensor Scheduling for Energy-Efficient Target Tracking in Sensor Networks
暂无分享,去创建一个
George Atia | Venugopal V. Veeravalli | Jason A. Fuemmeler | George K. Atia | V. Veeravalli | J. Fuemmeler
[1] G. Monahan. State of the Art—A Survey of Partially Observable Markov Decision Processes: Theory, Models, and Algorithms , 1982 .
[2] John W. Fisher,et al. Approximate Dynamic Programming for Communication-Constrained Sensor Network Management , 2007, IEEE Transactions on Signal Processing.
[3] Lihua Xie,et al. Multi-Sensor Scheduling for Reliable Target Tracking in Wireless Sensor Networks , 2006, 2006 6th International Conference on ITS Telecommunications.
[4] Leslie Pack Kaelbling,et al. Learning Policies for Partially Observable Environments: Scaling Up , 1997, ICML.
[5] Leslie Pack Kaelbling,et al. Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..
[6] Edwin K. P. Chong,et al. Sensor scheduling for target tracking: A Monte Carlo sampling approach , 2006, Digit. Signal Process..
[7] Darryl Morrell,et al. Nonmyopic Sensor Scheduling and its Efficient Implementation for Target Tracking Applications , 2006, EURASIP J. Adv. Signal Process..
[8] Stephen P. Boyd,et al. Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.
[9] Alfred O. Hero,et al. Sensor management using an active sensing approach , 2005, Signal Process..
[10] H. Vincent Poor,et al. An Introduction to Signal Detection and Estimation , 1994, Springer Texts in Electrical Engineering.
[11] H. Vincent Poor,et al. An introduction to signal detection and estimation (2nd ed.) , 1994 .
[12] Jesse Hoey,et al. Solving POMDPs with Continuous or Large Discrete Observation Spaces , 2005, IJCAI.
[13] D. Castañón. Approximate dynamic programming for sensor management , 1997, Proceedings of the 36th IEEE Conference on Decision and Control.
[14] Venugopal V. Veeravalli,et al. Smart Sleeping Policies for Energy Efficient Tracking in Sensor Networks , 2008, IEEE Transactions on Signal Processing.
[15] Michael L. Littman,et al. Incremental Pruning: A Simple, Fast, Exact Method for Partially Observable Markov Decision Processes , 1997, UAI.
[16] Joelle Pineau,et al. Point-based value iteration: An anytime algorithm for POMDPs , 2003, IJCAI.
[17] Alfred O. Hero,et al. A Bayesian method for integrated multitarget tracking and sensor management , 2003, Sixth International Conference of Information Fusion, 2003. Proceedings of the.
[18] Nicholas Roy,et al. Exponential Family PCA for Belief Compression in POMDPs , 2002, NIPS.
[19] Milos Hauskrecht,et al. Value-Function Approximations for Partially Observable Markov Decision Processes , 2000, J. Artif. Intell. Res..
[20] ZhaoFeng,et al. Collaborative in-network processing for target tracking , 2003 .
[21] Feng Zhao,et al. Collaborative In-Network Processing for Target Tracking , 2003, EURASIP J. Adv. Signal Process..
[22] George E. Monahan,et al. A Survey of Partially Observable Markov Decision Processes: Theory, Models, and Algorithms , 2007 .
[23] M. Kalandros. Covariance control for multisensor systems , 2002 .
[24] Feng Zhao,et al. Information-driven dynamic sensor collaboration , 2002, IEEE Signal Process. Mag..
[25] Nikos A. Vlassis,et al. Perseus: Randomized Point-based Value Iteration for POMDPs , 2005, J. Artif. Intell. Res..
[26] Alf Isaksson,et al. On sensor scheduling via information theoretic criteria , 1999, Proceedings of the 1999 American Control Conference (Cat. No. 99CH36251).
[27] Dimitri P. Bertsekas,et al. Dynamic Programming and Optimal Control, Two Volume Set , 1995 .
[28] Bernard C. Levy,et al. Principles of Signal Detection and Parameter Estimation , 2008 .
[29] Edward J. Sondik,et al. The optimal control of par-tially observable Markov processes , 1971 .
[30] Andrew W. Moore,et al. Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..
[31] Pascal Poupart,et al. Point-Based Value Iteration for Continuous POMDPs , 2006, J. Mach. Learn. Res..
[32] N. Zhang,et al. Algorithms for partially observable markov decision processes , 2001 .