Spatially Arranged Sparse Recurrent Neural Networks for Energy Efficient Associative Memory

The development of hardware neural networks, including neuromorphic hardware, has been accelerated over the past few years. However, it is challenging to operate very large-scale neural networks with low-power hardware devices, partly due to signal transmissions through a massive number of interconnections. Our aim is to deal with the issue of communication cost from an algorithmic viewpoint and study learning algorithms for energy-efficient information processing. Here, we consider two approaches to finding spatially arranged sparse recurrent neural networks with the high cost-performance ratio for associative memory. In the first approach following classical methods, we focus on sparse modular network structures inspired by biological brain networks and examine their storage capacity under an iterative learning rule. We show that incorporating long-range intermodule connections into purely modular networks can enhance the cost-performance ratio. In the second approach, we formulate for the first time an optimization problem where the network sparsity is maximized under the constraints imposed by a pattern embedding condition. We show that there is a tradeoff between the interconnection cost and the computational performance in the optimized networks. We demonstrate that the optimized networks can achieve a better cost-performance ratio compared with those considered in the first approach. We show the effectiveness of the optimization approach mainly using binary patterns and apply it also to gray-scale image restoration. Our results suggest that the presented approaches are useful in seeking more sparse and less costly connectivity of neural networks for the enhancement of energy efficiency in hardware neural networks.

[1]  Duncan J. Watts,et al.  Collective dynamics of ‘small-world’ networks , 1998, Nature.

[2]  Takashi Odagaki,et al.  Storage capacity and retrieval time of small-world neural networks. , 2007, Physical review. E, Statistical, nonlinear, and soft matter physics.

[3]  Lav R. Varshney,et al.  Structural Properties of the Caenorhabditis elegans Neuronal Network , 2009, PLoS Comput. Biol..

[4]  Ali A. Minai,et al.  Efficient associative memory using small-world architecture , 2001, Neurocomputing.

[5]  George B. Dantzig,et al.  Linear programming and extensions , 1965 .

[6]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[7]  Shimeng Yu,et al.  Synaptic electronics: materials, devices and applications , 2013, Nanotechnology.

[8]  Michael Menzinger,et al.  Topology and computational performance of attractor neural networks. , 2003, Physical review. E, Statistical, nonlinear, and soft matter physics.

[9]  E Gardner Optimal basins of attraction in randomly sparse neural network models , 1989 .

[10]  Alessandro Treves,et al.  Metastable states in asymmetrically diluted Hopfield networks , 1988 .

[11]  Jeferson Jacob Arenzon,et al.  Simulating highly diluted neural networks , 1994 .

[12]  J. Hopfield,et al.  Computing with neural circuits: a model. , 1986, Science.

[13]  Gorka Zamora-López,et al.  Cortical Hubs Form a Module for Multisensory Integration on Top of the Hierarchy of Cortical Networks , 2009, Front. Neuroinform..

[14]  Franck Vermet,et al.  Capacity of an associative memory model on random graph architectures , 2013, 1303.4542.

[15]  Anirvan M. Sengupta,et al.  Resource-efficient perceptron has sparse synaptic weight distribution , 2017, 2017 25th Signal Processing and Communications Applications Conference (SIU).

[16]  Toshiyuki Yamane,et al.  Regularity and randomness in modular network structures for neural associative memories , 2015, 2015 International Joint Conference on Neural Networks (IJCNN).

[17]  Wansheng Tang,et al.  A simple method for designing efficient small-world neural networks , 2010, Neural Networks.

[18]  Rodrigo Alvarez-Icaza,et al.  Neurogrid: A Mixed-Analog-Digital Multichip System for Large-Scale Neural Simulations , 2014, Proceedings of the IEEE.

[19]  Matthias Löwe,et al.  The Hopfield Model on a Sparse Erdös-Renyi Graph , 2011 .

[20]  Anton Bovier,et al.  Rigorous bounds on the storage capacity of the dilute Hopfield model , 1992 .

[21]  Opper,et al.  Learning of correlated patterns in spin-glass networks by local learning rules. , 1987, Physical review letters.

[22]  Dmitri B. Chklovskii,et al.  Wiring Optimization in Cortical Circuits , 2002, Neuron.

[23]  Neil Davey,et al.  High capacity associative memories and connection constraints , 2004, Connect. Sci..

[24]  Jonathan J. Hull,et al.  A Database for Handwritten Text Recognition Research , 1994, IEEE Trans. Pattern Anal. Mach. Intell..

[25]  Toshiyuki Yamane,et al.  Hopfield-Type Associative Memory with Sparse Modular Networks , 2014, ICONIP.

[26]  Olaf Sporns,et al.  Complex network measures of brain connectivity: Uses and interpretations , 2010, NeuroImage.

[27]  Jennifer Hasler,et al.  Finding a roadmap to achieve large neuromorphic hardware systems , 2013, Front. Neurosci..

[28]  L. da Fontoura Costa,et al.  Efficient Hopfield pattern recognition on a scale-free neural network , 2002, cond-mat/0212601.

[29]  Mark C. W. van Rossum,et al.  Energy Efficient Sparse Connectivity from Imbalanced Synaptic Plasticity Rules , 2015, PLoS Comput. Biol..

[30]  Aggelos K. Katsaggelos,et al.  Image restoration using a modified Hopfield network , 1992, IEEE Trans. Image Process..

[31]  Indranil Saha,et al.  journal homepage: www.elsevier.com/locate/neucom , 2022 .

[32]  J. Knott The organization of behavior: A neuropsychological theory , 1951 .

[33]  Akira Hirose,et al.  Complex-Valued Neural Networks , 2006, Studies in Computational Intelligence.

[34]  Jinde Cao,et al.  Topology influences performance in the associative memory neural networks , 2006 .

[35]  Balas K. Natarajan,et al.  Sparse Approximate Solutions to Linear Systems , 1995, SIAM J. Comput..

[36]  Armen Stepanyants,et al.  Effects of homeostatic constraints on associative memory storage and synaptic connectivity of cortical circuits , 2015, Front. Comput. Neurosci..

[37]  Fernando Morgado Dias,et al.  Artificial neural networks: a review of commercial hardware , 2004, Eng. Appl. Artif. Intell..

[38]  D. Chklovskii,et al.  Wiring optimization can relate neuronal structure and function. , 2006, Proceedings of the National Academy of Sciences of the United States of America.

[39]  O. Sporns,et al.  The economy of brain network organization , 2012, Nature Reviews Neuroscience.

[40]  John J. Hopfield,et al.  Neural networks and physical systems with emergent collective computational abilities , 1999 .

[41]  Yinyu Ye,et al.  A note on the complexity of Lp minimization , 2011, Math. Program..

[42]  K. Aihara,et al.  Complex-Valued Multistate Associative Memory With Nonlinear Multilevel Functions for Gray-Level Image Reconstruction , 2009, IEEE Transactions on Neural Networks.

[43]  E. Gardner,et al.  An Exactly Solvable Asymmetric Neural Network Model , 1987 .

[44]  Sompolinsky,et al.  Neural networks with nonlinear synapses and a static noise. , 1986, Physical review. A, General physics.

[45]  Toshiyuki Yamane,et al.  Wave-Based Neuromorphic Computing Framework for Brain-Like Energy Efficiency and Integration , 2016, IEEE Transactions on Nanotechnology.

[46]  Neil Davey,et al.  High Capacity Recurrent Associative Memories , 2002, Neurocomputing.

[47]  E. Gardner The space of interactions in neural network models , 1988 .

[48]  Michael A. Saunders,et al.  Atomic Decomposition by Basis Pursuit , 1998, SIAM J. Sci. Comput..

[49]  Andrew S. Cassidy,et al.  A million spiking-neuron integrated circuit with a scalable communication network and interface , 2014, Science.

[50]  Michael Elad,et al.  From Sparse Solutions of Systems of Equations to Sparse Modeling of Signals and Images , 2009, SIAM Rev..

[51]  Beom Jun Kim Performance of networks of artificial neurons: the role of clustering. , 2004, Physical review. E, Statistical, nonlinear, and soft matter physics.

[52]  Santosh S. Venkatesh,et al.  The capacity of the Hopfield associative memory , 1987, IEEE Trans. Inf. Theory.