Design Space Exploration of Dense and Sparse Mapping Schemes for RRAM Architectures

The impact of device and circuit-level effects in mixed-signal Resistive Random Access Memory (RRAM) accelerators typically manifest as performance degradation of Deep Learning (DL) algorithms, but the degree of impact varies based on algorithmic features. These include network architecture, capacity, weight distribution, and the type of inter-layer connections. Techniques are continuously emerging to efficiently train sparse neural networks, which may have activation sparsity, quantization, and memristive noise. In this paper, we present an extended Design Space Exploration (DSE) methodology to quantify the benefits and limitations of dense and sparse mapping schemes for a variety of network architectures. While sparsity of connectivity promotes less power consumption and is often optimized for extracting localized features, its performance on tiled RRAM arrays may be more susceptible to noise due to under-parameterization, when compared to dense mapping schemes. Moreover, we present a case study quantifying and formalizing the trade-offs of typical non-idealities introduced into 1-Transistor-1-Resistor (1T1R) tiled memristive architectures and the size of modular crossbar tiles using the CIFAR-10 dataset.

[1]  Yu Wang,et al.  Training low bitwidth convolutional neural network on RRAM , 2018, 2018 23rd Asia and South Pacific Design Automation Conference (ASP-DAC).

[2]  Yandong Luo,et al.  Impact of Read Disturb on Multilevel RRAM based Inference Engine: Experiments and Model Prediction , 2020, 2020 IEEE International Reliability Physics Symposium (IRPS).

[3]  Yu Wang,et al.  MErging the Interface: Power, area and accuracy co-optimization for RRAM crossbar-based mixed-signal computing system , 2015, 2015 52nd ACM/EDAC/IEEE Design Automation Conference (DAC).

[4]  Dominique Drouin,et al.  In‐Memory Vector‐Matrix Multiplication in Monolithic Complementary Metal–Oxide–Semiconductor‐Memristor Integrated Circuits: Design Choices, Challenges, and Perspectives , 2020, Adv. Intell. Syst..

[5]  Wenbo Zhao,et al.  SME: ReRAM-based Sparse-Multiplication-Engine to Squeeze-Out Bit Sparsity of Neural Network , 2021, 2021 IEEE 39th International Conference on Computer Design (ICCD).

[6]  Zhuowen Tu,et al.  Aggregated Residual Transformations for Deep Neural Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Tao Zhang,et al.  Overcoming the challenges of crossbar resistive memory architectures , 2015, 2015 IEEE 21st International Symposium on High Performance Computer Architecture (HPCA).

[8]  Mark Sandler,et al.  MobileNetV2: Inverted Residuals and Linear Bottlenecks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[9]  Cong Xu,et al.  Design trade-offs for high density cross-point resistive memory , 2012, ISLPED '12.

[10]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[11]  Naveen Verma,et al.  Neural Network Training With Stochastic Hardware Models and Software Abstractions , 2021, IEEE Transactions on Circuits and Systems I: Regular Papers.

[12]  Chih-Yang Lin,et al.  Complementary Metal‐Oxide Semiconductor and Memristive Hardware for Neuromorphic Computing , 2020, Adv. Intell. Syst..

[13]  Kyeong-Sik Min,et al.  Comparative Study on Quantization-Aware Training of Memristor Crossbars for Reducing Inference Power of Neural Networks at The Edge , 2021, 2021 International Joint Conference on Neural Networks (IJCNN).

[14]  Xiaochen Peng,et al.  DNN+NeuroSim V2.0: An End-to-End Benchmarking Framework for Compute-in-Memory Accelerators for On-Chip Training , 2020, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[15]  Yiran Chen,et al.  Accelerator-friendly neural-network training: Learning variations and defects in RRAM crossbar , 2017, Design, Automation & Test in Europe Conference & Exhibition (DATE), 2017.

[16]  Mohammed E. Fouda,et al.  IR-QNN Framework: An IR Drop-Aware Offline Training of Quantized Crossbar Arrays , 2020, IEEE Access.

[17]  Bin Gao,et al.  Fully hardware-implemented memristor convolutional neural network , 2020, Nature.

[18]  Olga Krestinskaya,et al.  Automating Analogue AI Chip Design with Genetic Search , 2020, Adv. Intell. Syst..

[19]  Tayfun Gokmen,et al.  A Flexible and Fast PyTorch Toolkit for Simulating Training and Inference on Analog Crossbar Arrays , 2021, 2021 IEEE 3rd International Conference on Artificial Intelligence Circuits and Systems (AICAS).

[20]  Mostafa Rahimi Azghadi,et al.  MemTorch: An Open-source Simulation Framework for Memristive Deep Learning Systems , 2020, ArXiv.

[21]  Mohammed A. Zidan,et al.  A Crossbar-Based In-Memory Computing Architecture , 2020, IEEE Transactions on Circuits and Systems I: Regular Papers.

[22]  Meng-Fan Chang,et al.  RRAM-based coprocessors for deep learning , 2020 .

[23]  Erich Elsen,et al.  The State of Sparsity in Deep Neural Networks , 2019, ArXiv.

[24]  Yiran Chen,et al.  A quantization-aware regularized learning method in multilevel memristor-based neuromorphic computing system , 2017, 2017 IEEE 6th Non-Volatile Memory Systems and Applications Symposium (NVMSA).

[25]  Norbert Wehn,et al.  RRAMSpec: A Design Space Exploration Framework for High Density Resistive RAM , 2019, SAMOS.

[26]  Bernabe Linares-Barranco,et al.  Hardware Implementation of Deep Network Accelerators Towards Healthcare and Biomedical Applications , 2020, IEEE Transactions on Biomedical Circuits and Systems.

[27]  Mohammed Bennamoun,et al.  Training Spiking Neural Networks Using Lessons From Deep Learning , 2021, ArXiv.

[28]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[29]  D. Ielmini,et al.  Empirical metal-oxide RRAM device endurance and retention model for deep learning simulations , 2021, Semiconductor Science and Technology.

[30]  Natalia Gimelshein,et al.  PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.

[31]  Mostafa Rahimi Azghadi,et al.  Variation-aware Binarized Memristive Networks , 2019, 2019 26th IEEE International Conference on Electronics, Circuits and Systems (ICECS).

[32]  Cong Xu,et al.  Design implications of memristor-based RRAM cross-point structures , 2011, 2011 Design, Automation & Test in Europe.