Computing-In-Memory Neural Network Accelerators for Safety-Critical Systems: Can Small Device Variations Be Disastrous?

Computing-in-Memory (CiM) architectures based on emerging non-volatile memory (NVM) devices have demonstrated great potential for deep neural network (DNN) acceleration thanks to their high energy efficiency. However, NVM devices suffer from various non- idealities, especially device-to-device variations due to fabrication defects and cycle-to-cycle variations due to the stochastic behavior of devices. As such, the DNN weights actually mapped to NVM devices could deviate significantly from the expected values, leading to large performance degradation. To address this issue, most existing works focus on maximizing average performance under device variations. This objective would work well for general-purpose scenarios. But for safety-critical applications, the worst-case performance must also be considered. Unfortunately, this has been rarely explored in the literature. In this work, we formulate the problem of determining the worst-case performance of CiM DNN accelerators under the impact of device variations. We further propose a method to effectively find the specific combination of device variation in the high-dimensional space that leads to the worst-case performance. We find that even with very small device variations, the accuracy of a DNN can drop drastically, causing concerns when deploying CiM accelerators in safety-critical applications. Finally, we show that surprisingly none of the existing methods used to enhance average DNN performance in CiM accelerators are very effective when extended to enhance the worst-case performance, and further research down the road is needed to address this problem.

[1]  2021 Design, Automation & Test in Europe Conference & Exhibition (DATE) , 2023, 2023 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[2]  Yoonmyung Lee,et al.  Variation-Tolerant and Low R-Ratio Compute-in-Memory ReRAM Macro With Capacitive Ternary MAC Operation , 2022, IEEE Transactions on Circuits and Systems I: Regular Papers.

[3]  Xuehai Qian,et al.  Variational Quantum Pulse Learning , 2022, 2022 IEEE International Conference on Quantum Computing and Engineering (QCE).

[4]  Yiyu Shi,et al.  A Semi-Decoupled Approach to Fast and Optimal Hardware-Software Co-Design of Neural Accelerators , 2022, ArXiv.

[5]  Zhilu Wang,et al.  Efficient Global Robustness Certification of Neural Networks via Interleaving Twin-Network Encoding , 2022, 2022 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[6]  Yiyu Shi,et al.  The larger the fairer?: small neural networks can achieve fairness for edge devices , 2022, DAC.

[7]  X. Hu,et al.  SWIM: selective write-verify for computing-in-memory neural accelerators , 2022, DAC.

[8]  Krishnendu Chakrabarty,et al.  Pruning of Deep Neural Networks for Fault-Tolerant Memristor-based Accelerators , 2021, 2021 58th ACM/IEEE Design Automation Conference (DAC).

[9]  Lee-Sup Kim,et al.  Fault-free: A Fault-resilient Deep Neural Network Accelerator based on Realistic ReRAM Devices , 2021, 2021 58th ACM/IEEE Design Automation Conference (DAC).

[10]  Ulf Schlichtmann,et al.  Bayesian Inference Based Robust Computing on Memristor Crossbar , 2021, 2021 58th ACM/IEEE Design Automation Conference (DAC).

[11]  Nanyang Ye,et al.  BayesFT: Bayesian Optimization for Fault Tolerant Neural Network Architecture , 2021, 2021 58th ACM/IEEE Design Automation Conference (DAC).

[12]  Yiyu Shi,et al.  RADARS: Memory Efficient Reinforcement Learning Aided Differentiable Neural Architecture Search , 2021, 2022 27th Asia and South Pacific Design Automation Conference (ASP-DAC).

[13]  Shanshi Huang,et al.  A 40nm RRAM Compute-in-Memory Macro Featuring On-Chip Write-Verify and Offset-Cancelling ADC References , 2021, ESSCIRC 2021 - IEEE 47th European Solid State Circuits Conference (ESSCIRC).

[14]  Jinjun Xiong,et al.  Can Noise on Qubits Be Learned in Quantum Neural Network? A Case Study on QuantumFlow (Invited Paper) , 2021, 2021 IEEE/ACM International Conference On Computer Aided Design (ICCAD).

[15]  Jinjun Xiong,et al.  Exploration of Quantum Neural Architecture by Mixing Quantum Neuron Designs: (Invited Paper) , 2021, 2021 IEEE/ACM International Conference On Computer Aided Design (ICCAD).

[16]  Yiyu Shi,et al.  Uncertainty Modeling of Emerging Device based Computing-in-Memory Neural Accelerators with Application to Neural Architecture Search , 2021, 2021 26th Asia and South Pacific Design Automation Conference (ASP-DAC).

[17]  Shimeng Yu,et al.  Two-step write–verify scheme and impact of the read noise in multilevel RRAM-based inference engine , 2020, Semiconductor Science and Technology.

[18]  Yisen Wang,et al.  Adversarial Weight Perturbation Helps Robust Generalization , 2020, NeurIPS.

[19]  Brendan Juba,et al.  More Accurate Learning of k-DNF Reference Classes , 2020, AAAI.

[20]  Songwei Pei,et al.  On Improving Fault Tolerance of Memristor Crossbar Based Neural Network Designs by Target Sparsifying , 2020, 2020 Design, Automation & Test in Europe Conference & Exhibition (DATE).

[21]  Meng Li,et al.  Co-Exploration of Neural Architectures and Heterogeneous ASIC Accelerator Designs Targeting Multiple Tasks , 2020, 2020 57th ACM/IEEE Design Automation Conference (DAC).

[22]  Xiaochen Peng,et al.  DNN+NeuroSim: An End-to-End Benchmarking Framework for Compute-in-Memory Accelerators with Versatile Device Technologies , 2019, 2019 IEEE International Electron Devices Meeting (IEDM).

[23]  X. Hu,et al.  Device-Circuit-Architecture Co-Exploration for Computing-in-Memory Neural Accelerators , 2019, IEEE Transactions on Computers.

[24]  Masanori Hashimoto,et al.  When Single Event Upset Meets Deep Neural Networks: Observations, Explorations, and Remedies , 2019, 2020 25th Asia and South Pacific Design Automation Conference (ASP-DAC).

[25]  Jie Lin,et al.  Noise Injection Adaption: End-to-End ReRAM Crossbar Non-ideal Effect Adaption for Neural Network Mapping , 2019, 2019 56th ACM/IEEE Design Automation Conference (DAC).

[26]  Tao Liu,et al.  A Fault-Tolerant Neural Network Architecture , 2019, 2019 56th ACM/IEEE Design Automation Conference (DAC).

[27]  Heng-Yuan Lee,et al.  NV-BNN: An Accurate Deep Convolutional Neural Network Based on Binary STT-MRAM for Adaptive AI Edge , 2019, 2019 56th ACM/IEEE Design Automation Conference (DAC).

[28]  Engin Ipek,et al.  Making Memristive Neural Network Accelerators Reliable , 2018, 2018 IEEE International Symposium on High Performance Computer Architecture (HPCA).

[29]  Vivienne Sze,et al.  Efficient Processing of Deep Neural Networks: A Tutorial and Survey , 2017, Proceedings of the IEEE.

[30]  Miao Hu,et al.  ISAAC: A Convolutional Neural Network Accelerator with In-Situ Analog Arithmetic in Crossbars , 2016, 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA).

[31]  Vivienne Sze,et al.  Eyeriss: A Spatial Architecture for Energy-Efficient Dataflow for Convolutional Neural Networks , 2016, 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA).

[32]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[33]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[34]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[35]  L. Deng,et al.  The MNIST Database of Handwritten Digit Images for Machine Learning Research [Best of the Web] , 2012, IEEE Signal Processing Magazine.

[36]  Fei-Fei Li,et al.  ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[37]  Pin-Yu Chen,et al.  Formalizing Generalization and Adversarial Robustness of Neural Networks to Weight Perturbations , 2021, NeurIPS.

[38]  Ya Le,et al.  Tiny ImageNet Visual Recognition Challenge , 2015 .

[39]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[40]  Simon Haykin,et al.  GradientBased Learning Applied to Document Recognition , 2001 .

[41]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.