Energy-Constrained Distributed Learning and Classification by Exploiting Relative Relevance of Sensors’ Data

We consider the problem of communicating data from energy-constrained distributed sensors. To reduce energy requirements, we go beyond the source reconstruction problem classically addressed, and focus on the problem where the recipient wants to perform supervised learning and classification on the data received from the sensors. Restricting our attention to a noiseless communication setting under simplistic Gaussian source assumptions, we study supervised learning and classification under total energy limitations. The energy constraints are modeled in two ways: 1) a linear scaling and 2) an exponential scaling of energy with number of bits used for compression at sensors. We first assume that the underlying parameters for Gaussian distributions have already been learned, and obtain (with linear scaling, reverse-waterfilling-type) strategies for allocating energy, and thus, bits, across different sensors under these two models. Intuitively, these strategies allocate larger rates and energies to sensors that are more “relevant” for the classification goal. These strategies are used to obtain an achievable bound on the tradeoff between energy and error-probability (classification risk). We then provide an algorithm for learning the distribution-parameters of the sensor-data under energy constraints to arrive at high-reliability energy-allocation strategies, while enabling the energy-allocation algorithm to backtrack when the underlying distributions change, or when there is noise in sensed data that can push the algorithm toward a local minimum. Finally, we provide numerical results on energy-savings for classification of simulated data as well as neural data acquired from electrocorticography (ECoG) experiments.

[1]  Thomas M. Cover,et al.  A Proof of the Data Compression Theorem of Slepian and Wolf for Ergodic Sources , 1971 .

[2]  Anant Sahai,et al.  Towards a Communication-Theoretic Understanding of System-Level Power Consumption , 2010, IEEE Journal on Selected Areas in Communications.

[3]  Bernhard Schölkopf,et al.  Support vector channel selection in BCI , 2004, IEEE Transactions on Biomedical Engineering.

[4]  David L. Neuhoff,et al.  Quantization , 2022, IEEE Trans. Inf. Theory.

[5]  P. F. Panter,et al.  Quantization distortion in pulse-count modulation with nonuniform spacing of levels , 1951, Proceedings of the IRE.

[6]  J. Lafferty,et al.  Quantized Nonparametric Estimation , 2015 .

[7]  Robert D. Nowak,et al.  Compressive distilled sensing: Sparse recovery using adaptivity in compressive measurements , 2009, 2009 Conference Record of the Forty-Third Asilomar Conference on Signals, Systems and Computers.

[8]  Urbashi Mitra,et al.  KNOWME: An Energy-Efficient Multimodal Body Area Network for Physical Activity Monitoring , 2012, TECS.

[9]  R. Nowak,et al.  Active Sensing and Learning , 2011 .

[10]  Stuart Barber,et al.  All of Statistics: a Concise Course in Statistical Inference , 2005 .

[11]  Robert D. Nowak,et al.  Active Clustering: Robust and Efficient Hierarchical Clustering using Adaptively Selected Similarities , 2011, AISTATS.

[12]  Rudolf Ahlswede,et al.  Hypothesis testing with communication constraints , 1986, IEEE Trans. Inf. Theory.

[13]  Robert Tibshirani,et al.  The Elements of Statistical Learning: Data Mining, Inference, and Prediction, 2nd Edition , 2001, Springer Series in Statistics.

[14]  S. P. Lloyd,et al.  Least squares quantization in PCM , 1982, IEEE Trans. Inf. Theory.

[15]  Sergio VerdÂ,et al.  Statistical Inference Under Multiterminal Data Compression , 2000 .

[16]  Pulkit Grover Ultra-high-density EEG: How many bits of resolution do the electrodes need? , 2015, 2015 49th Asilomar Conference on Signals, Systems and Computers.

[17]  Robert D. Nowak,et al.  Compressed Sensing Vs. Active Learning , 2006, 2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings.

[18]  Pramod K. Varshney,et al.  Energy-Aware Sensor Selection in Field Reconstruction , 2014, IEEE Signal Processing Letters.

[19]  L. Schuchman Dither Signals and Their Effect on Quantization Noise , 1964 .

[20]  Xin Li,et al.  Using relative-relevance of data pieces for efficient communication, with an application to Neural data acquisition , 2014, 2014 52nd Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[21]  J. Fermaglich Electric Fields of the Brain: The Neurophysics of EEG , 1982 .

[22]  Urbashi Mitra,et al.  Boundary Estimation in Sensor Networks: Theory and Methods , 2003, IPSN.

[23]  Katalin Marton,et al.  Error exponent for source coding with a fidelity criterion , 1974, IEEE Trans. Inf. Theory.

[24]  David Blaauw,et al.  Energy-efficient dot product computation using a switched analog circuit architecture , 2014, 2014 IEEE/ACM International Symposium on Low Power Electronics and Design (ISLPED).

[25]  T. Hinterberger,et al.  Automated EEG feature selection for brain computer interfaces , 2003, First International IEEE EMBS Conference on Neural Engineering, 2003. Conference Proceedings..

[26]  Xin Li,et al.  Low-power hardware implementation of movement decoding for brain computer interface with reduced-resolution discrete cosine transform , 2014, 2014 36th Annual International Conference of the IEEE Engineering in Medicine and Biology Society.

[27]  Robert D. Nowak,et al.  Faster Rates in Regression via Active Learning , 2005, NIPS.

[28]  Tian Lan,et al.  Salient EEG Channel Selection in Brain Computer Interfaces by Mutual Information Maximization , 2005, 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference.

[29]  Monica A. Perez,et al.  Neural interface technology for rehabilitation: exploiting and promoting neuroplasticity. , 2010, Physical medicine and rehabilitation clinics of North America.

[30]  Yijun Wang,et al.  Common Spatial Pattern Method for Channel Selelction in Motor Imagery Based Brain-computer Interface , 2005, 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference.

[31]  Toby Berger,et al.  The CEO problem [multiterminal source coding] , 1996, IEEE Trans. Inf. Theory.

[32]  R. Zamir,et al.  On lattice quantization noise , 1994, Proceedings of 1994 IEEE International Symposium on Information Theory.

[33]  Robert D. Nowak,et al.  Distilled Sensing: Adaptive Sampling for Sparse Detection and Estimation , 2010, IEEE Transactions on Information Theory.

[34]  Zhen Zhang,et al.  On the CEO problem , 1994, Proceedings of 1994 IEEE International Symposium on Information Theory.