Activity-dependent neuron model for noise resistance

Abstract Activity-dependent plasticity plays an important role in biological neural network learning. Unlike the backpropagation-based learning in artificial neural networks that depends on supervised signals, biological neurons adjust themselves using their own historical behaviors as a clue to benefit information processing. Inspired by this biological neural mechanism, this paper proposes a novel neuron model, named Activity-Dependent Neuron (ADN). The basic idea of ADN is to add a gate in the neuron model to control its signal conduction capability, that is, the gate will facilitate transmission of important signals while suppress trivial signals. This idea can be achieved by self-tuning of activation behaviors. We also develop an iterative training algorithm, so that the ADNs can be smoothly incorporated into deep neural networks to jointly learn with the network weights. Experimental results found that the ADNs can efficiently improve the noise-resistant capability. Compared with the state-of-the-art, it is more robust to unforeseen noises. It does not need noise data for training.

[1]  F. Auer,et al.  Evidence for Myelin Sheath Remodeling in the CNS Revealed by In Vivo Imaging , 2018, Current Biology.

[2]  R. Douglas Fields,et al.  A new mechanism of nervous system plasticity: activity-dependent myelination , 2015, Nature Reviews Neuroscience.

[3]  Samuel Schacher,et al.  The less things change, the more they are different: contributions of long-term synaptic plasticity and homeostasis to memory , 2014, Learning & memory.

[4]  D. Hubel,et al.  Receptive fields of single neurones in the cat's striate cortex , 1959, The Journal of physiology.

[5]  Geoffrey E. Hinton,et al.  Dynamic Routing Between Capsules , 2017, NIPS.

[6]  Joaquín J. Torres,et al.  Inverse stochastic resonance in networks of spiking neurons , 2017, PLoS Comput. Biol..

[7]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[8]  Marian Tsanov,et al.  Intrinsic, Light-Independent and Visual Activity-Dependent Mechanisms Cooperate in the Shaping of the Field Response in Rat Visual Cortex , 2007, The Journal of Neuroscience.

[9]  Jorge Nocedal,et al.  A Limited Memory Algorithm for Bound Constrained Optimization , 1995, SIAM J. Sci. Comput..

[10]  Y. LeCun,et al.  Learning methods for generic object recognition with invariance to pose and lighting , 2004, Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004..

[11]  D. Hassabis,et al.  Neuroscience-Inspired Artificial Intelligence , 2017, Neuron.

[12]  Karunesh Ganguly,et al.  Activity-Dependent Neural Plasticity from Bench to Bedside , 2013, Neuron.

[13]  Andrew Y. Ng,et al.  Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .

[14]  Lawrence D. Jackel,et al.  Backpropagation Applied to Handwritten Zip Code Recognition , 1989, Neural Computation.

[15]  Bhaskara Marthi,et al.  A generative vision model that trains with high data efficiency and breaks text-based CAPTCHAs , 2017, Science.

[16]  Dipankar Das,et al.  Enhanced SenticNet with Affective Labels for Concept-Based Opinion Mining , 2013, IEEE Intelligent Systems.

[17]  Razvan Pascanu,et al.  Vector-based navigation using grid-like representations in artificial agents , 2018, Nature.

[18]  Zhaohui Wu,et al.  Jointly Learning Network Connections and Link Weights in Spiking Neural Networks , 2018, IJCAI.

[19]  Mahmut Ozer,et al.  Impact of hybrid neural network on the early diagnosis of diabetic retinopathy disease from video-oculography signals , 2018, Chaos, Solitons & Fractals.

[20]  Geoffrey E. Hinton,et al.  Matrix capsules with EM routing , 2018, ICLR.

[21]  D. Buonomano,et al.  A Novel Learning Rule for Long-Term Plasticity of Short-Term Synaptic Plasticity Enhances Temporal Processing , 2011, Front. Integr. Neurosci..

[22]  A. Faisal,et al.  Noise in the nervous system , 2008, Nature Reviews Neuroscience.

[23]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[24]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[25]  G. Turrigiano The Self-Tuning Neuron: Synaptic Scaling of Excitatory Synapses , 2008, Cell.

[26]  M. Baudry,et al.  Properties and Mechanisms of Long-Term Synaptic Plasticity in the Mammalian Brain: Relationships to Learning and Memory , 1995, Neurobiology of Learning and Memory.

[27]  Gang Pan,et al.  Rapid Decoding of Hand Gestures in Electrocorticography Using Recurrent Neural Networks , 2018, Front. Neurosci..

[28]  J. J. Torres,et al.  Effects of dynamic synapses on noise-delayed response latency of a single neuron. , 2015, Physical review. E, Statistical, nonlinear, and soft matter physics.

[29]  Mahmut Ozer,et al.  Double inverse stochastic resonance with dynamic synapses. , 2016, Physical review. E.

[30]  Geoffrey E. Hinton,et al.  Learning to combine foveal glimpses with a third-order Boltzmann machine , 2010, NIPS.