Source-Channel Communication in Sensor Networks

Sensors acquire data, and communicate this to an interested party. The arising coding problem is often split into two parts: First, the sensors compress their respective acquired signals, potentially applying the concepts of distributed source coding. Then, they communicate the compressed version to the interested party, the goal being not to make any errors. This coding paradigm is inspired by Shannon's separation theorem for point-to-point communication, but it leads to suboptimal performance in general network topologies. The optimal performance for the general case is not known. In this paper, we propose an alternative coding paradigm based on joint source-channel coding. This coding paradigm permits to determine the optimal performance for a class of sensor networks, and shows how to achieve it. For sensor networks outside this class, we argue that the goal of the coding system could be to approach our condition for optimal performance as closely as possible. This is supported by examples for which our coding paradigm significantly outperforms the traditional separation-based coding paradigm. In particular, for a Gaussian example considered in this paper, the distortion of the best coding scheme according to the separation paradigm decreases like 1/logM, while for our coding paradigm, it decreases like 1/M, where M is the total number of sensors.

[1]  Toby Berger,et al.  The quadratic Gaussian CEO problem , 1997, IEEE Trans. Inf. Theory.

[2]  Cindy C. Parman,et al.  To code, or not to code? , 2003, The Journal of oncology management : the official journal of the American College of Oncology Administrators.

[3]  Michael Gastpar,et al.  To code, or not to code: lossy source-channel communication revisited , 2003, IEEE Trans. Inf. Theory.

[4]  Masoud Salehi,et al.  Multiple access channels with arbitrarily correlated sources , 1980, IEEE Trans. Inf. Theory.

[5]  Michael Gastpar,et al.  On the capacity of large Gaussian relay networks , 2005, IEEE Transactions on Information Theory.

[6]  Toby Berger,et al.  The CEO problem [multiterminal source coding] , 1996, IEEE Trans. Inf. Theory.

[7]  Zhen Zhang,et al.  On the CEO problem , 1994, Proceedings of 1994 IEEE International Symposium on Information Theory.

[8]  Giuseppe Longo,et al.  The information theory approach to communications , 1977 .

[9]  Aaron D. Wyner,et al.  The rate-distortion function for source coding with side information at the decoder , 1976, IEEE Trans. Inf. Theory.

[10]  J. Wolfowitz The rate distortion function for source coding with side information at the decoder , 1979 .

[11]  Yasutada Oohama,et al.  The Rate-Distortion Function for the Quadratic Gaussian CEO Problem , 1998, IEEE Trans. Inf. Theory.

[12]  Michael Gastpar,et al.  On the capacity of wireless networks: the relay case , 2002, Proceedings.Twenty-First Annual Joint Conference of the IEEE Computer and Communications Societies.

[13]  Jack K. Wolf,et al.  Noiseless coding of correlated information sources , 1973, IEEE Trans. Inf. Theory.

[14]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[15]  Thomas J. Goblick,et al.  Theoretical limitations on the transmission of data from analog sources , 1965, IEEE Trans. Inf. Theory.

[16]  Michael Gastpar,et al.  To code or not to code , 2000, 2000 IEEE International Symposium on Information Theory (Cat. No.00CH37060).