Low area overhead in-situ training approach for memristor-based classifier

We propose combination of "dropout" and "Manhattan Rule" training algorithms for memristive crossbar neural networks to reduce circuit area overhead of in-situ training. Using accurate phenomenological model of memristive devices, we show that such combination allows achieving 0.7% misclassification rate on the MNIST benchmark, which is comparable to the best reported results. At the same time, the considered training approach allows reducing the size of memory circuits, the largest area overhead component, which is required to store intermediate weight adjustments during training, by as much as 40% at 16% longer training time as compared to the baseline crossbar circuit compatible "Manhattan Rule" training. The further reduction of the memory circuit area overhead is possible but at the expense of inferior classification performance.

[1]  Patrice Y. Simard,et al.  Best practices for convolutional neural networks applied to visual document analysis , 2003, Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings..

[2]  Farnood Merrikh-Bayat,et al.  Training and operation of an integrated neuromorphic network based on metal-oxide memristors , 2014, Nature.

[3]  Fabien Alibart,et al.  Pattern classification by memristive crossbar circuits using ex situ and in situ training , 2013, Nature Communications.

[4]  Yong Zhang,et al.  A digital neuromorphic VLSI architecture with memristor crossbar synaptic array for machine learning , 2012, 2012 IEEE International SOC Conference.

[5]  D. Strukov,et al.  Phenomenological modeling of memristive devices , 2014, 1406.4219.

[6]  Dmitri B. Strukov,et al.  Nanotechnology: Smart connections , 2011, Nature.

[7]  Konstantin K. Likharev,et al.  CrossNets: Neuromorphic Hybrid CMOS/Nanoelectronic Networks , 2011 .

[8]  Farnood Merrikh-Bayat,et al.  Efficient training algorithms for neural networks based on memristive crossbar circuits , 2015, 2015 International Joint Conference on Neural Networks (IJCNN).

[9]  Dmitri Strukov,et al.  Manhattan rule training for memristive crossbar circuit pattern classifiers , 2015, 2015 IEEE 9th International Symposium on Intelligent Signal Processing (WISP) Proceedings.

[10]  Berin Martini,et al.  NeuFlow: A runtime reconfigurable dataflow processor for vision , 2011, CVPR 2011 WORKSHOPS.

[11]  Konstantin K. Likharev,et al.  Defect‐tolerant nanoelectronic pattern classifiers , 2007, Int. J. Circuit Theory Appl..

[12]  Steven R. Young,et al.  A 1 TOPS/W Analog Deep Machine-Learning Engine With Floating-Gate Storage in 0.13 µm CMOS , 2014, IEEE Journal of Solid-State Circuits.

[13]  Nitish Srivastava,et al.  Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.

[14]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.