Variational Probability Flow for Biologically Plausible Training of Deep Neural Networks
暂无分享,去创建一个
[1] Geoffrey E. Hinton,et al. Reducing the Dimensionality of Data with Neural Networks , 2006, Science.
[2] S.-I. Amari,et al. Neural theory of association and concept-formation , 1977, Biological Cybernetics.
[3] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[4] Xiaohui Xie,et al. Spike-based Learning Rules and Stabilization of Persistent Neural Activity , 1999, NIPS.
[5] Yoshua Bengio,et al. STDP-Compatible Approximation of Backpropagation in an Energy-Based Model , 2017, Neural Computation.
[6] Yoshua Bengio,et al. Towards Biologically Plausible Deep Learning , 2015, ArXiv.
[7] Geoffrey E. Hinton,et al. A View of the Em Algorithm that Justifies Incremental, Sparse, and other Variants , 1998, Learning in Graphical Models.
[8] Yoshua Bengio,et al. Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation , 2016, Front. Comput. Neurosci..
[9] Alex Graves,et al. Decoupled Neural Interfaces using Synthetic Gradients , 2016, ICML.
[10] Wulfram Gerstner,et al. Variational Learning for Recurrent Spiking Networks , 2011, NIPS.
[11] Yg,et al. Dropout as a Bayesian Approximation : Insights and Applications , 2015 .
[12] G. Bi,et al. Synaptic Modifications in Cultured Hippocampal Neurons: Dependence on Spike Timing, Synaptic Strength, and Postsynaptic Cell Type , 1998, The Journal of Neuroscience.
[13] Pascal Vincent,et al. Parallel Tempering for Training of Restricted Boltzmann Machines , 2010 .
[14] Rémi Monasson,et al. Emergence of Compositional Representations in Restricted Boltzmann Machines , 2016, Physical review letters.
[15] Jascha Sohl-Dickstein,et al. Minimum Probability Flow Learning , 2009, ICML.
[16] Geoffrey E. Hinton,et al. Deep Boltzmann Machines , 2009, AISTATS.
[17] Yoshua Bengio,et al. Better Mixing via Deep Representations , 2012, ICML.
[18] Wulfram Gerstner,et al. A neuronal learning rule for sub-millisecond temporal coding , 1996, Nature.
[19] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[20] Joel Z. Leibo,et al. How Important Is Weight Symmetry in Backpropagation? , 2015, AAAI.
[21] Yoshua Bengio,et al. Deep Generative Stochastic Networks Trainable by Backprop , 2013, ICML.
[22] John Salvatier,et al. Theano: A Python framework for fast computation of mathematical expressions , 2016, ArXiv.
[23] Geoffrey E. Hinton. Training Products of Experts by Minimizing Contrastive Divergence , 2002, Neural Computation.
[24] Jascha Sohl-Dickstein,et al. A new method for parameter estimation in probabilistic models: Minimum probability flow , 2011, Physical review letters.
[25] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[26] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[27] Hugo Larochelle,et al. Efficient Learning of Deep Boltzmann Machines , 2010, AISTATS.
[28] Geoffrey E. Hinton,et al. A Learning Algorithm for Boltzmann Machines , 1985, Cogn. Sci..