Analog Weights in ReRAM DNN Accelerators

Artificial neural networks have become ubiquitous in modern life, which has triggered the emergence of a new class of application specific integrated circuits for their acceleration. ReRAM-based accelerators have gained significant traction due to their ability to leverage in-memory computations. In a crossbar structure, they can perform multiply-and-accumulate operations more efficiently than standard CMOS logic. By virtue of being resistive switches, ReRAM switches can only reliably store one of two states. This is a severe limitation on the range of values in a computational kernel. This paper presents a novel scheme in alleviating the single-bit-per-device restriction by exploiting frequency dependence of v-i plane hysteresis, and assigning kernel information not only to the device conductance but also partially distributing it to the frequency of a time-varying input.We show this approach reduces average power consumption for a single crossbar convolution by up to a factor of ×16 for an unsigned 8-bit input image, where each convolutional process consumes a worst-case of 1.1mW, and reduces area by a factor of ×8, without reducing accuracy to the level of binarized neural networks. This presents a massive saving in computing cost when there are many simultaneous in-situ multiply-and-accumulate processes occurring across different crossbars.

[1]  Kyoung-Rok Cho,et al.  Maximization of Crossbar Array Memory Using Fundamental Memristor Theory , 2017, IEEE Transactions on Circuits and Systems II: Express Briefs.

[2]  Ran El-Yaniv,et al.  Binarized Neural Networks , 2016, ArXiv.

[3]  Tyrone Fernando,et al.  Analysis and generation of chaos using compositely connected coupled memristors. , 2018, Chaos.

[4]  Farnood Merrikh-Bayat,et al.  Training and operation of an integrated neuromorphic network based on metal-oxide memristors , 2014, Nature.

[5]  Jiaming Zhang,et al.  Analogue signal and image processing with large memristor crossbars , 2017, Nature Electronics.

[6]  L.O. Chua,et al.  Memristive devices and systems , 1976, Proceedings of the IEEE.

[7]  Tyrone Fernando,et al.  Modelling and characterization of dynamic behavior of coupled memristor circuits , 2016, 2016 IEEE International Symposium on Circuits and Systems (ISCAS).

[8]  Kyoungrok Cho,et al.  Nano-Programmable Logics Based on Double-Layer Anti-Facing Memristors. , 2019, Journal of nanoscience and nanotechnology.

[9]  Kyoung-Rok Cho,et al.  Formulation and Implementation of Nonlinear Integral Equations to Model Neural Dynamics Within the Vertebrate Retina , 2018, Int. J. Neural Syst..

[10]  Nicolangelo Iannella,et al.  Signal Flow Platform for Mapping and Simulation of Vertebrate Retina for Sensor Systems , 2016, IEEE Sensors Journal.

[11]  Chris Yakopcic,et al.  Exploring the design space of specialized multicore neural processors , 2013, The 2013 International Joint Conference on Neural Networks (IJCNN).

[12]  Yong Zhang,et al.  A digital neuromorphic VLSI architecture with memristor crossbar synaptic array for machine learning , 2012, 2012 IEEE International SOC Conference.

[13]  Yiran Chen,et al.  PipeLayer: A Pipelined ReRAM-Based Accelerator for Deep Learning , 2017, 2017 IEEE International Symposium on High Performance Computer Architecture (HPCA).

[14]  Hao Jiang,et al.  RENO: A high-efficient reconfigurable neuromorphic computing accelerator design , 2015, 2015 52nd ACM/EDAC/IEEE Design Automation Conference (DAC).

[15]  G. Cauwenberghs,et al.  Memristor-based neural networks: Synaptic versus neuronal stochasticity , 2016 .

[16]  M. A. Nugent,et al.  AHaH Computing–From Metastable Switches to Attractors to Machine Learning , 2014, PloS one.

[17]  Miao Hu,et al.  ISAAC: A Convolutional Neural Network Accelerator with In-Situ Analog Arithmetic in Crossbars , 2016, 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA).