Forward Signal Propagation Learning
暂无分享,去创建一个
[1] Garrick Orchard,et al. Advancing Neuromorphic Computing With Loihi: A Survey of Results and Outlook , 2021, Proceedings of the IEEE.
[2] Edith Beigné,et al. Spiking Neural Networks Hardware Implementations and Challenges , 2019, ACM J. Emerg. Technol. Comput. Syst..
[3] Michael Eickenberg,et al. Decoupled Greedy Learning of CNNs , 2019, ICML.
[4] Arild Nøkland,et al. Training Neural Networks with Local Error Signals , 2019, ICML.
[5] Jacques Kaiser,et al. Synaptic Plasticity Dynamics for Deep Continuous Local Learning (DECOLLE) , 2018, Frontiers in Neuroscience.
[6] Yoshua Bengio,et al. Extending the Framework of Equilibrium Propagation to General Dynamics , 2018, ICLR.
[7] Shih-Chii Liu,et al. Conversion of Continuous-Valued Deep Networks to Efficient Event-Driven Networks for Image Classification , 2017, Front. Neurosci..
[8] Yoshua Bengio,et al. Equivalence of Equilibrium Propagation and Recurrent Backpropagation , 2017, Neural Computation.
[9] Takayoshi Yoshimura,et al. A Parallel Forward-Backward Propagation Learning Scheme for Auto-Encoders , 2017, ICONIP.
[10] Chaitali Chakrabarti,et al. Algorithm and hardware design of discrete-time spiking neural networks based on back propagation with binary activations , 2017, 2017 IEEE Biomedical Circuits and Systems Conference (BioCAS).
[11] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[12] Max Jaderberg,et al. Understanding Synthetic Gradients and Decoupled Neural Interfaces , 2017, ICML.
[13] Somnath Paul,et al. Event-Driven Random Back-Propagation: Enabling Neuromorphic Deep Learning Machines , 2016, Front. Neurosci..
[14] Colin J. Akerman,et al. Random synaptic feedback weights support error backpropagation for deep learning , 2016, Nature Communications.
[15] Timothy P Lillicrap,et al. Towards deep learning with segregated dendrites , 2016, eLife.
[16] Arild Nøkland,et al. Direct Feedback Alignment Provides Learning in Deep Neural Networks , 2016, NIPS.
[17] Tobi Delbrück,et al. Training Deep Spiking Neural Networks Using Backpropagation , 2016, Front. Neurosci..
[18] Alex Graves,et al. Decoupled Neural Interfaces using Synthetic Gradients , 2016, ICML.
[19] Frank Hutter,et al. SGDR: Stochastic Gradient Descent with Warm Restarts , 2016, ICLR.
[20] Konrad P. Körding,et al. Toward an Integration of Deep Learning and Neuroscience , 2016, bioRxiv.
[21] Yoshua Bengio,et al. Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation , 2016, Front. Comput. Neurosci..
[22] Joel Z. Leibo,et al. How Important Is Weight Symmetry in Backpropagation? , 2015, AAAI.
[23] Matthew Cook,et al. Fast-classifying, high-accuracy spiking deep networks through weight and threshold balancing , 2015, 2015 International Joint Conference on Neural Networks (IJCNN).
[24] Yongqiang Cao,et al. Spiking Deep Convolutional Neural Networks for Energy-Efficient Object Recognition , 2015, International Journal of Computer Vision.
[25] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[26] Yoshua Bengio,et al. Difference Target Propagation , 2014, ECML/PKDD.
[27] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[28] Yoshua Bengio,et al. How Auto-Encoders Could Provide Credit Assignment in Deep Networks via Target Propagation , 2014, ArXiv.
[29] Mark Horowitz,et al. 1.1 Computing's energy problem (and what we can do about it) , 2014, 2014 IEEE International Solid-State Circuits Conference Digest of Technical Papers (ISSCC).
[30] Stefan Schliebs,et al. Span: Spike Pattern Association Neuron for Learning Spatio-Temporal Spike Patterns , 2012, Int. J. Neural Syst..
[31] Naohiro Fukumura,et al. A forward-propagation learning rule for neural inverse models using a method of recursive least squares , 2005, Systems and Computers in Japan.
[32] Naohiro Fukumura,et al. A Forward-Propagation Rule for Acquiring Neural Inverse Models Using a RLS Algorithm , 2004, ICONIP.
[33] Xiaohui Xie,et al. Equivalence of Backpropagation and Contrastive Hebbian Learning in a Layered Network , 2003, Neural Computation.
[34] Nihar R. Mahapatra,et al. The processor-memory bottleneck: problems and solutions , 1999, CROS.
[35] Kotaro Hirasawa,et al. Forward propagation universal learning network , 1996, Proceedings of International Conference on Neural Networks (ICNN'96).
[36] A. P. Heinz. Pipelined neural tree learning by error forward-propagation , 1995, Proceedings of ICNN'95 - International Conference on Neural Networks.
[37] Francis Crick,et al. The recent excitement about neural networks , 1989, Nature.
[38] Stephen Grossberg,et al. Competitive Learning: From Interactive Activation to Adaptive Resonance , 1987, Cogn. Sci..
[39] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[40] J J Hopfield,et al. Neurons with graded response have collective computational properties like those of two-state neurons. , 1984, Proceedings of the National Academy of Sciences of the United States of America.
[41] John W. Backus,et al. Can programming be liberated from the von Neumann style?: a functional style and its algebra of programs , 1978, CACM.
[42] Andrew L. Maas. Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .
[43] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[44] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[45] Yann LeCun,et al. The mnist database of handwritten digits , 2005 .
[46] Ronald J. Williams,et al. Gradient-based learning algorithms for recurrent connectionist networks , 1990 .