Analog VLSI Stochastic Perturbative Learning Architectures

We present analog VLSI neuromorphic architectures fora general class of learning tasks, which include supervised learning,reinforcement learning, and temporal difference learning. Thepresented architectures are parallel, cellular, sparse in globalinterconnects, distributed in representation, and robust to noiseand mismatches in the implementation. They use a parallel stochasticperturbation technique to estimate the effect of weight changeson network outputs, rather than calculating derivatives basedon a model of the network. This “model-free” technique avoidserrors due to mismatches in the physical implementation of thenetwork, and more generally allows to train networks of whichthe exact characteristics and structure are not known. With additionalmechanisms of reinforcement learning, networks of fairly generalstructure are trained effectively from an arbitrarily suppliedreward signal. No prior assumptions are required on the structureof the network nor on the specifics of the desired network response.

[1]  Gert Cauwenberghs,et al.  A Fast Stochastic Error-Descent Algorithm for Supervised Learning and Optimization , 1992, NIPS.

[2]  T. Sacktor The Synaptic Organization of the Brain (3rd Ed.) , 1991 .

[3]  P. Werbos,et al.  Beyond Regression : "New Tools for Prediction and Analysis in the Behavioral Sciences , 1974 .

[4]  Fernando J. Pineda,et al.  Mean-Field Theory for Batched TD() , 1997, Neural Computation.

[5]  E. Kandel,et al.  A cellular mechanism of classical conditioning in Aplysia: activity-dependent amplification of presynaptic facilitation. , 1983, Science.

[6]  Ron Meir,et al.  A Parallel Gradient Descent Method for Learning in Analog VLSI Neural Networks , 1992, NIPS.

[7]  Richard S. Sutton,et al.  A Menu of Designs for Reinforcement Learning Over Time , 1995 .

[8]  Gert Cauwenberghs A micropower CMOS algorithmic A/D/A converter , 1995 .

[9]  E. Vittoz,et al.  Analog Storage of Adjustable Synaptic Weights , 1991 .

[10]  Richard S. Sutton,et al.  Neuronlike adaptive elements that can solve difficult learning control problems , 1983, IEEE Transactions on Systems, Man, and Cybernetics.

[11]  S. Kelso,et al.  Differential conditioning of associative synaptic enhancement in hippocampal brain slices. , 1986, Science.

[12]  Gert Cauwenberghs Analog VLSI long-term dynamic storage , 1996, 1996 IEEE International Symposium on Circuits and Systems. Circuits and Systems Connecting the World. ISCAS 96.

[13]  Marwan A. Jabri,et al.  Summed Weight Neuron Perturbation: An O(N) Improvement Over Weight Perturbation , 1992, NIPS.

[14]  J. Spall A Stochastic Approximation Technique for Generating Maximum Likelihood Parameter Estimates , 1987, 1987 American Control Conference.

[15]  S. Grossberg,et al.  Neural dynamics of attentionally modulated Pavlovian conditioning: blocking, interstimulus interval, and secondary reinforcement. , 1987, Applied optics.

[16]  S. Grossberg A neural model of attention, reinforcement and discrimination learning. , 1975, International review of neurobiology.

[17]  Marwan A. Jabri,et al.  Weight perturbation: an optimal architecture and learning technique for analog VLSI feedforward and recurrent multilayer networks , 1992, IEEE Trans. Neural Networks.

[18]  Carver Mead,et al.  Analog VLSI and neural systems , 1989 .

[19]  Gert Cauwenberghs A Learning Analog Neural Network Chip with Continuous-Time Recurrent Dynamics , 1993, NIPS.

[20]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[21]  Edgar Sanchez-Sinencio,et al.  Artificial Neural Networks: Paradigms, Applications, and Hardware Implementations , 1994 .

[22]  Richard S. Sutton,et al.  Neural networks for control , 1990 .

[23]  Gert Cauwenberghs,et al.  Fault-tolerant dynamic multilevel storage in analog VLSI , 1994 .

[24]  Kurt W. Fleischer,et al.  Analog VLSI Implementation of Gradient Descent , 1992, NIPS.

[25]  Gabor C. Temes,et al.  Oversampling Delta Sigma Data Converters , 1991 .

[26]  Peter Dayan,et al.  Bee foraging in uncertain environments using predictive hebbian learning , 1995, Nature.

[27]  Carver A. Mead,et al.  A single-transistor silicon synapse , 1996 .

[28]  J. Fellrath,et al.  CMOS analog integrated circuits based on weak inversion operations , 1977 .

[29]  Andreas G. Andreou,et al.  Current-mode subthreshold MOS circuits for analog VLSI neural systems , 1991, IEEE Trans. Neural Networks.

[30]  Mohammed Ismail,et al.  Analog VLSI Implementation of Neural Systems , 2011, The Kluwer International Series in Engineering and Computer Science.

[31]  M. A. Styblinski,et al.  Experiments in nonconvex optimization: Stochastic approximation with function smoothing and simulated annealing , 1990, Neural Networks.

[32]  Gert Cauwenberghs,et al.  An analog VLSI recurrent neural network learning a continuous-time trajectory , 1996, IEEE Trans. Neural Networks.

[33]  A. Hodgkin,et al.  Currents carried by sodium and potassium ions through the membrane of the giant axon of Loligo , 1952, The Journal of physiology.

[34]  Thomas Kailath,et al.  Model-free distributed learning , 1990, IEEE Trans. Neural Networks.

[35]  Harold J. Kushner,et al.  wchastic. approximation methods for constrained and unconstrained systems , 1978 .

[36]  Carver A. Mead,et al.  Neuromorphic electronic systems , 1990, Proc. IEEE.

[37]  Gert Cauwenberghs,et al.  Analysis and verification of an analog VLSI incremental outer-product learning system , 1992, IEEE Trans. Neural Networks.

[38]  Terrence J. Sejnowski,et al.  The Computational Brain , 1996, Artif. Intell..

[39]  W. Precht The synaptic organization of the brain G.M. Shepherd, Oxford University Press (1975). 364 pp., £3.80 (paperback) , 1976, Neuroscience.

[40]  Eduardo D. Sontag,et al.  Neural Networks for Control , 1993 .

[41]  Paul J. Werbos,et al.  The roots of backpropagation , 1994 .

[42]  Tamio Shimizu,et al.  A Stochastic Approximation Method for Optimization Problems , 1969, Journal of the ACM.