Attractor Dynamics in Feedforward Neural Networks
暂无分享,去创建一个
[1] N. Metropolis,et al. Equation of State Calculations by Fast Computing Machines , 1953, Resonance.
[2] J J Hopfield,et al. Neural networks and physical systems with emergent collective computational abilities. , 1982, Proceedings of the National Academy of Sciences of the United States of America.
[3] Stephen Grossberg,et al. Absolute stability of global pattern formation and parallel memory storage by competitive neural networks , 1983, IEEE Transactions on Systems, Man, and Cybernetics.
[4] J J Hopfield,et al. Neurons with graded response have collective computational properties like those of two-state neurons. , 1984, Proceedings of the National Academy of Sciences of the United States of America.
[5] Donald Geman,et al. Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images , 1984, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[6] Geoffrey E. Hinton,et al. A Learning Algorithm for Boltzmann Machines , 1985, Cogn. Sci..
[7] J. Hopfield,et al. Computing with neural circuits: a model. , 1986, Science.
[8] C Koch,et al. Analog "neuronal" networks in early vision. , 1986, Proceedings of the National Academy of Sciences of the United States of America.
[9] Carsten Peterson,et al. A Mean Field Theory Learning Algorithm for Neural Networks , 1987, Complex Syst..
[10] Judea Pearl,et al. Probabilistic reasoning in intelligent systems , 1988 .
[11] Judea Pearl,et al. Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.
[12] Gregory F. Cooper,et al. The Computational Complexity of Probabilistic Inference Using Bayesian Belief Networks , 1990, Artif. Intell..
[13] Michael I. Jordan,et al. Forward Models: Supervised Learning with a Distal Teacher , 1992, Cogn. Sci..
[14] Radford M. Neal. Connectionist Learning of Belief Networks , 1992, Artif. Intell..
[15] Jung-Hsien Chiang,et al. Training neural pattern classifiers with a mean field theory learning algorithm , 1992 .
[16] Jenq-Neng Hwang,et al. Iterative inversion of neural networks and its application to adaptive control , 1992, IEEE Trans. Neural Networks.
[17] C. Galland. The limitations of deterministic Boltzmann machine learning , 1993 .
[18] Stuart J. Russell,et al. Adaptive Probabilistic Networks , 1994 .
[19] Wray L. Buntine. Operations for Learning with Graphical Models , 1994, J. Artif. Intell. Res..
[20] Geoffrey E. Hinton,et al. The Helmholtz Machine , 1995, Neural Computation.
[21] Geoffrey E. Hinton,et al. The "wake-sleep" algorithm for unsupervised neural networks. , 1995, Science.
[22] Hill,et al. Annealed Theories of Learning , 1995 .
[23] Jong-Hoon Oh,et al. Neural networks : the statistical mechanics perspective : proceedings of the CTP-PBSRI Joint Workshop on Theoretical Physics, POSTECH, Pohang, Korea, 2-4 February 95 , 1995 .
[24] Brendan J. Frey,et al. Does the Wake-sleep Algorithm Produce Good Density Estimators? , 1995, NIPS.
[25] Michael I. Jordan,et al. Mean Field Theory for Sigmoid Belief Networks , 1996, J. Artif. Intell. Res..
[26] Terrence J. Sejnowski,et al. Bayesian Unsupervised Learning of Higher Order Structure , 1996, NIPS.
[27] Jeffrey C. Lagarias,et al. Minimax and Hamiltonian Dynamics of Excitatory-Inhibitory Networks , 1997, NIPS.
[28] Geoffrey E. Hinton,et al. Generative models for discovering sparse distributed representations. , 1997, Philosophical transactions of the Royal Society of London. Series B, Biological sciences.
[29] Michael I. Jordan. Graphical Models , 2003 .
[30] Geoffrey E. Hinton. Training Products of Experts by Minimizing Contrastive Divergence , 2002, Neural Computation.