On the impact of smart sensor approximations on the accuracy of machine learning tasks

Smart sensors present in ubiquitous Internet of Things (IoT) devices often obtain high energy efficiency by carefully tuning how the sensing, the analog to digital (A/D) conversion and the digital serial transmission are implemented. Such tuning involves approximations, i.e. alterations of the sensed signals that can positively affect energy consumption in various ways. However, for many IoT applications, approximations may have an impact on the quality of the produced output, for example on the classification accuracy of a Machine Learning (ML) model. While the impact of approximations on ML algorithms is widely studied, previous works have focused mostly on processing approximations. In this work, in contrast, we analyze how the signal alterations imposed by smart sensors impact the accuracy of ML classifiers. We focus in particular on data alterations introduced in the serial transmission from a smart sensor to a processor, although our considerations can also be extended to other sources of approximation, such as A/D conversion. Results on several types of models and on two different datasets show that ML algorithms are quite resilient to the alterations produced by smart sensors, and that the serial transmission energy can be reduced by up to 70% without a significant impact on classification accuracy. Moreover, we also show that, contrarily to expectations, the two generic approximation families identified in our work yield similar accuracy losses.

[1]  Fabio Roli,et al.  Wild Patterns: Ten Years After the Rise of Adversarial Machine Learning , 2018, CCS.

[2]  Anand Raghunathan,et al.  AXSERBUS: A quality-configurable approximate serial bus for energy-efficient sensing , 2017, 2017 IEEE/ACM International Symposium on Low Power Electronics and Design (ISLPED).

[3]  Marimuthu Palaniswami,et al.  Energy-efficient data acquisition by adaptive sampling for wireless sensor networks , 2008, IWCMC.

[4]  Timo Aila,et al.  Pruning Convolutional Neural Networks for Resource Efficient Inference , 2016, ICLR.

[5]  Marian Verhelst,et al.  Energy-efficient ConvNets through approximate computing , 2016, 2016 IEEE Winter Conference on Applications of Computer Vision (WACV).

[6]  Vivienne Sze,et al.  Efficient Processing of Deep Neural Networks: A Tutorial and Survey , 2017, Proceedings of the IEEE.

[7]  Soheil Ghiasi,et al.  Hardware-oriented Approximation of Convolutional Neural Networks , 2016, ArXiv.

[8]  Sachin S. Talathi,et al.  Fixed Point Quantization of Deep Convolutional Networks , 2015, ICML.

[9]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[10]  Sameep Mehta,et al.  A Survey on Resilient Machine Learning , 2017, ArXiv.

[11]  Arnab Raha,et al.  Energy-efficient system design for IoT devices , 2016, 2016 21st Asia and South Pacific Design Automation Conference (ASP-DAC).

[12]  Daniela Micucci,et al.  UniMiB SHAR: a new dataset for human activity recognition using acceleration data from smartphones , 2016, ArXiv.

[13]  Ling Jian,et al.  A Noise-Resilient Online Learning Algorithm for Scene Classification , 2018, Remote. Sens..

[14]  Sergey Ioffe,et al.  Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[15]  S. Chakrabartty,et al.  A dynamic reconfigurable A/D converter for sensor applications , 2005, IEEE Sensors, 2005..

[16]  Enrico Macii,et al.  Approximate differential encoding for energy-efficient serial communication , 2016, 2016 International Great Lakes Symposium on VLSI (GLSVLSI).

[17]  Felix Wortmann,et al.  Internet of Things , 2015, Business & Information Systems Engineering.

[18]  G. G. Stokes "J." , 1890, The New Yale Book of Quotations.

[19]  Lina J. Karam,et al.  Understanding how image quality affects deep neural networks , 2016, 2016 Eighth International Conference on Quality of Multimedia Experience (QoMEX).

[20]  Enrico Macii,et al.  Approximate Energy-Efficient Encoding for Serial Interfaces , 2017, ACM Trans. Design Autom. Electr. Syst..

[21]  Ran El-Yaniv,et al.  Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations , 2016, J. Mach. Learn. Res..

[22]  Luca Rigazio,et al.  Towards Deep Neural Network Architectures Robust to Adversarial Examples , 2014, ICLR.

[23]  Enrico Macii,et al.  Battery-aware Design Exploration of Scheduling Policies for Multi-sensor Devices , 2018, ACM Great Lakes Symposium on VLSI.

[24]  Xukan Ran,et al.  Deep Learning With Edge Computing: A Review , 2019, Proceedings of the IEEE.

[25]  Enrico Macii,et al.  Energy-Efficient Digital Processing via Approximate Computing , 2016 .

[26]  Enrico Macii,et al.  Zero-Transition Serial Encoding for Image Sensors , 2017, IEEE Sensors Journal.

[27]  Martin C. Rinard,et al.  Error-Efficient Computing Systems , 2017, Found. Trends Electron. Des. Autom..

[28]  Enrico Macii,et al.  Dynamic Bit-width Reconfiguration for Energy-Efficient Deep Learning Hardware , 2018, ISLPED.

[29]  Muriel Medard,et al.  Energy-efficient Time-Stampless Adaptive Nonuniform Sampling , 2011, 2011 IEEE SENSORS Proceedings.

[30]  Anand Raghunathan,et al.  AxBA: An Approximate Bus Architecture Framework , 2018, 2018 IEEE/ACM International Conference on Computer-Aided Design (ICCAD).

[31]  A.P. Chandrakasan,et al.  A 10-pJ/instruction, 4-MIPS micropower DSP for sensor applications , 2008, 2008 IEEE Asian Solid-State Circuits Conference.

[32]  Martin C. Rinard,et al.  A Hardware Platform for Efficient Multi-Modal Sensing with Adaptive Approximation , 2018, ArXiv.

[33]  Di Wu,et al.  Approximate Hardware Techniques for Energy-Quality Scaling Across the System , 2020, 2020 International Conference on Electronics, Information, and Communication (ICEIC).

[34]  Hui Liu,et al.  A 9-b 40-MSample/s reconfigurable pipeline analog-to-digital converter , 2002 .

[35]  Qiang Xu,et al.  ApproxANN: An approximate computing framework for artificial neural network , 2015, 2015 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[36]  Luca Benini,et al.  XNOR Neural Engine: A Hardware Accelerator IP for 21.6-fJ/op Binary Neural Network Inference , 2018, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[37]  Ilaria Scarabottolo,et al.  A spectrum-based adaptive sampling algorithm for smart sensing , 2017, 2017 IEEE SmartWorld, Ubiquitous Intelligence & Computing, Advanced & Trusted Computed, Scalable Computing & Communications, Cloud & Big Data Computing, Internet of People and Smart City Innovation (SmartWorld/SCALCOM/UIC/ATC/CBDCom/IOP/SCI).

[38]  Kamal El-Sankary,et al.  Practical Considerations for Accuracy Evaluation in Sensor-Based Machine Learning and Deep Learning , 2019, Sensors.

[39]  Tsuyoshi Murata,et al.  {m , 1934, ACML.

[40]  Ali Farhadi,et al.  XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks , 2016, ECCV.

[41]  Martin C. Rinard,et al.  Reducing serial I/O power in error-tolerant applications by efficient lossy encoding , 2016, 2016 53nd ACM/EDAC/IEEE Design Automation Conference (DAC).

[42]  Hoi-Jun Yoo,et al.  SILENT: serialized low energy transmission coding for on-chip interconnection networks , 2004, IEEE/ACM International Conference on Computer Aided Design, 2004. ICCAD-2004..

[43]  Thomas G. Dietterich,et al.  Benchmarking Neural Network Robustness to Common Corruptions and Perturbations , 2018, ICLR.

[44]  Enrico Macii,et al.  Serial T0: Approximate bus encoding for energy-efficient transmission of sensor signals , 2016, 2016 53nd ACM/EDAC/IEEE Design Automation Conference (DAC).

[45]  Martin Rinard,et al.  Warp: A Hardware Platform for Efficient Multimodal Sensing With Adaptive Approximation , 2020, IEEE Micro.

[46]  Mark Sandler,et al.  MobileNetV2: Inverted Residuals and Linear Bottlenecks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[47]  Song Han,et al.  Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.

[48]  Sparsh Mittal,et al.  A Survey of Techniques for Approximate Computing , 2016, ACM Comput. Surv..