Deep associative neural network for associative memory based on unsupervised representation learning

This paper presents a deep associative neural network (DANN) based on unsupervised representation learning for associative memory. In brain, the knowledge is learnt by associating different types of sensory data, such as image and voice. The associative memory models which imitate such a learning process have been studied for decades but with simpler architectures they fail to deal with large scale complex data as compared with deep neural networks. Therefore, we define a deep architecture consisting of a perception layer and hierarchical propagation layers. To learn the network parameters, we define a probabilistic model for the whole network inspired from unsupervised representation learning models. The model is optimized by a modified contrastive divergence algorithm with a novel iterated sampling process. After training, given a new data or corrupted data, the correct label or corrupted part is associated by the network. The DANN is able to achieve many machine learning problems, including not only classification, but also depicting the data given a label and recovering corrupted images. Experiments on MNIST digits and CIFAR-10 datasets demonstrate the learning capability of the proposed DANN.

[1]  Sumio Hosaka,et al.  Associative memory realized by a reconfigurable memristive Hopfield neural network , 2015, Nature Communications.

[2]  Dhabaleswar K. Panda,et al.  S-Caffe: Co-designing MPI Runtimes and Caffe for Scalable Deep Learning on Modern GPU Clusters , 2017, PPoPP.

[3]  Jürgen Schmidhuber,et al.  Learning Precise Timing with LSTM Recurrent Networks , 2003, J. Mach. Learn. Res..

[4]  Haibo He,et al.  Spatio–Temporal Memories for Machine Learning: A Long-Term Memory Organization , 2009, IEEE Transactions on Neural Networks.

[5]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[6]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Pascal Vincent,et al.  Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion , 2010, J. Mach. Learn. Res..

[8]  Yoshua Bengio,et al.  Greedy Layer-Wise Training of Deep Networks , 2006, NIPS.

[9]  David D. Vogel Auto-associative memory produced by disinhibition in a sparsely connected network , 1998, Neural Networks.

[10]  Jürgen Schmidhuber,et al.  Training Very Deep Networks , 2015, NIPS.

[11]  Derek C. Rose,et al.  Deep Machine Learning - A New Frontier in Artificial Intelligence Research [Research Frontier] , 2010, IEEE Computational Intelligence Magazine.

[12]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[13]  Yee Whye Teh,et al.  A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.

[14]  Maoguo Gong,et al.  Neuron Learning Machine for Representation Learning , 2017, AAAI.

[15]  John J. Hopfield,et al.  Dense Associative Memory for Pattern Recognition , 2016, NIPS.

[16]  Masaki Kobayashi Decomposition of Rotor Hopfield Neural Networks Using Complex Numbers , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[17]  Lipo Wang Heteroassociations of spatio-temporal sequences with the bidirectional associative memory , 2000, IEEE Trans. Neural Networks Learn. Syst..

[18]  Geoffrey E. Hinton Training Products of Experts by Minimizing Contrastive Divergence , 2002, Neural Computation.

[19]  Chun-Xia Zhang,et al.  A sparse-response deep belief network based on rate distortion theory , 2014, Pattern Recognit..

[20]  Eduard Kuriscak,et al.  Biological context of Hebb learning in artificial neural networks, a review , 2015, Neurocomputing.

[21]  Yoshua Bengio,et al.  Deep Sparse Rectifier Neural Networks , 2011, AISTATS.

[22]  Honglak Lee,et al.  Sparse deep belief net model for visual area V2 , 2007, NIPS.

[23]  Maoguo Gong,et al.  Modeling Hebb Learning Rule for Unsupervised Learning , 2017, IJCAI.

[24]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[25]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[26]  Haibo He,et al.  A Hierarchical Self-organizing Associative Memory for Machine Learning , 2007, ISNN.

[27]  Günther Palm,et al.  Neural associative memories and sparse coding , 2013, Neural Networks.

[28]  Maoguo Gong,et al.  A Multiobjective Sparse Feature Learning Model for Deep Neural Networks , 2015, IEEE Transactions on Neural Networks and Learning Systems.

[29]  Maoguo Gong,et al.  A Deep Convolutional Coupling Network for Change Detection Based on Heterogeneous Optical and Radar Images , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[30]  Pascal Vincent,et al.  Representation Learning: A Review and New Perspectives , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[31]  J J Hopfield,et al.  Neural networks and physical systems with emergent collective computational abilities. , 1982, Proceedings of the National Academy of Sciences of the United States of America.

[32]  Marc'Aurelio Ranzato,et al.  Sparse Feature Learning for Deep Belief Networks , 2007, NIPS.

[33]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).