暂无分享,去创建一个
[1] E. Adrian,et al. The impulses produced by sensory nerve‐endings , 1926 .
[2] David Reitter,et al. Learning to Adapt by Minimizing Discrepancy , 2017, ArXiv.
[3] Peter Elias,et al. Predictive coding-I , 1955, IRE Trans. Inf. Theory.
[4] Colin J. Akerman,et al. Random synaptic feedback weights support error backpropagation for deep learning , 2016, Nature Communications.
[5] Daniel Kifer,et al. Unifying Adversarial Training Algorithms with Data Gradient Regularization , 2017, Neural Computation.
[6] Razvan Pascanu,et al. On the difficulty of training recurrent neural networks , 2012, ICML.
[7] Marc'Aurelio Ranzato,et al. Fast Inference in Sparse Coding Algorithms with Applications to Object Recognition , 2010, ArXiv.
[8] A. Clark. Whatever next? Predictive brains, situated agents, and the future of cognitive science. , 2013, The Behavioral and brain sciences.
[9] David Reitter,et al. Online Semi-Supervised Learning with Deep Hybrid Boltzmann Machines and Denoising Autoencoders , 2015, ArXiv.
[10] Geoffrey E. Hinton,et al. Learning Representations by Recirculation , 1987, NIPS.
[11] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[12] Ji Feng,et al. Multi-Layered Gradient Boosting Decision Trees , 2018, NeurIPS.
[13] Stephen Grossberg,et al. Competitive Learning: From Interactive Activation to Adaptive Resonance , 1987, Cogn. Sci..
[14] James A. Anderson,et al. Neurocomputing: Foundations of Research , 1988 .
[15] Frank Moss,et al. Noise in human muscle spindles , 1996, Nature.
[16] Yoshua Bengio,et al. Difference Target Propagation , 2014, ECML/PKDD.
[17] Rahul Sarpeshkar,et al. Analog Versus Digital: Extrapolating from Electronics to Neurobiology , 1998, Neural Computation.
[18] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[19] David Reitter,et al. Online Learning of Deep Hybrid Architectures for Semi-supervised Categorization , 2015, ECML/PKDD.
[20] Joachim M. Buhmann,et al. Kickback Cuts Backprop's Red-Tape: Biologically Plausible Credit Assignment in Neural Networks , 2014, AAAI.
[21] Rajesh P. N. Rao,et al. Predictive Coding , 2019, A Blueprint for the Hard Problem of Consciousness.
[22] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[23] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[24] Rob R. de Ruyter van Steveninck,et al. The metabolic cost of neural information , 1998, Nature Neuroscience.
[25] G. J. Tomko,et al. Neuronal variability: non-stationary responses to identical visual stimuli. , 1974, Brain research.
[26] H. Dertinger,et al. Stochastic resonance as a possible mechanism of amplification of weak electric signals in living cells. , 1994, Bioelectromagnetics.
[27] Rafal Bogacz,et al. An Approximation of the Error Backpropagation Algorithm in a Predictive Coding Network with Local Hebbian Synaptic Plasticity , 2017, Neural Computation.
[28] Rajesh P. N. Rao,et al. Dynamic Model of Visual Recognition Predicts Neural Response Properties in the Visual Cortex , 1997, Neural Computation.
[29] L. Pinneo. On noise in the nervous system. , 1966, Psychological review.
[30] S. Grossberg,et al. How does a brain build a cognitive code? , 1980, Psychological review.
[31] W. Newsome,et al. The Variable Discharge of Cortical Neurons: Implications for Connectivity, Computation, and Information Coding , 1998, The Journal of Neuroscience.
[32] Arild Nøkland,et al. Direct Feedback Alignment Provides Learning in Deep Neural Networks , 2016, NIPS.
[33] Andrea Hasenstaub,et al. Barrages of Synaptic Activity Control the Gain and Sensitivity of Cortical Neurons , 2003, The Journal of Neuroscience.
[34] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[35] Geoffrey E. Hinton,et al. Visualizing Data using t-SNE , 2008 .
[36] Yoshua Bengio,et al. Greedy Layer-Wise Training of Deep Networks , 2006, NIPS.
[37] Zhuowen Tu,et al. Deeply-Supervised Nets , 2014, AISTATS.
[38] David J. Field,et al. Sparse coding with an overcomplete basis set: A strategy employed by V1? , 1997, Vision Research.
[39] J. Movshon,et al. The statistical reliability of signals in single neurons in cat and monkey visual cortex , 1983, Vision Research.
[40] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[41] Yoshua Bengio,et al. Towards Biologically Plausible Deep Learning , 2015, ArXiv.
[42] Alex Graves,et al. Decoupled Neural Interfaces using Synthetic Gradients , 2016, ICML.
[43] Yoshua Bengio,et al. Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation , 2016, Front. Comput. Neurosci..
[44] Moshe Bar,et al. Predictive Feedback and Conscious Visual Experience , 2012, Front. Psychology.
[45] Rajesh P. N. Rao,et al. Predictive coding in the visual cortex: a functional interpretation of some extra-classical receptive-field effects. , 1999 .
[46] Daniel Kifer,et al. Conducting Credit Assignment by Aligning Local Representations , 2018, 1803.01834.
[47] Joel Z. Leibo,et al. How Important Is Weight Symmetry in Backpropagation? , 2015, AAAI.
[48] Randall C. O'Reilly,et al. Biologically Plausible Error-Driven Learning Using Local Activation Differences: The Generalized Recirculation Algorithm , 1996, Neural Computation.
[49] Jun Li,et al. Predictive Coding Machine for Compressed Sensing and Image Denoising , 2018, AAAI.
[50] Daniel Cownden,et al. Random feedback weights support learning in deep neural networks , 2014, ArXiv.