Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion
暂无分享,去创建一个
Pascal Vincent | Yoshua Bengio | Hugo Larochelle | Pierre-Antoine Manzagol | Isabelle Lajoie | Yoshua Bengio | H. Larochelle | Pierre-Antoine Manzagol | Pascal Vincent | Isabelle Lajoie
[1] D. Hubel,et al. Receptive fields of single neurones in the cat's striate cortex , 1959, The Journal of physiology.
[2] J. Besag. Statistical Analysis of Non-Lattice Data , 1975 .
[3] J J Hopfield,et al. Neural networks and physical systems with emergent collective computational abilities. , 1982, Proceedings of the National Academy of Sciences of the United States of America.
[4] Paul Smolensky,et al. Information processing in dynamical systems: foundations of harmony theory , 1986 .
[5] Johan Håstad,et al. Almost optimal lower bounds for small depth circuits , 1986, STOC '86.
[6] James L. McClelland,et al. Parallel distributed processing: explorations in the microstructure of cognition, vol. 1: foundations , 1986 .
[7] Yann LeCun. PhD thesis: Modeles connexionnistes de l'apprentissage (connectionist learning models) , 1987 .
[8] Yann LeCun,et al. Memoires associatives distribuees: Une comparaison (Distributed associative memories: A comparison) , 1987 .
[9] Jay S. Patel,et al. Factors influencing learning by backpropagation , 1988, IEEE 1988 International Conference on Neural Networks.
[10] Richard T. Scalettar,et al. Emergence of grandmother memory in feed forward networks: learning with noise and forgetfulness , 1988 .
[11] Ralph Linsker,et al. An Application of the Principle of Maximum Information Preservation to Linear Systems , 1988, NIPS.
[12] Kurt Hornik,et al. Neural networks and principal component analysis: Learning from examples without local minima , 1989, Neural Networks.
[13] Geoffrey E. Hinton. Connectionist Learning Procedures , 1989, Artif. Intell..
[14] Lawrence D. Jackel,et al. Backpropagation Applied to Handwritten Zip Code Recognition , 1989, Neural Computation.
[15] Jocelyn Sietsma,et al. Creating artificial neural networks that generalize , 1991, Neural Networks.
[16] Yann LeCun,et al. Tangent Prop - A Formalism for Specifying Selected Invariances in an Adaptive Network , 1991, NIPS.
[17] Petri Koistinen,et al. Using additive noise in back-propagation training , 1992, IEEE Trans. Neural Networks.
[18] T. Poggio,et al. Recognition and Structure from one 2D Model View: Observations on Prototypes, Object Classes and Symmetries , 1992 .
[19] G. Kane. Parallel Distributed Processing: Explorations in the Microstructure of Cognition, vol 1: Foundations, vol 2: Psychological and Biological Models , 1994 .
[20] Terrence J. Sejnowski,et al. An Information-Maximization Approach to Blind Separation and Blind Deconvolution , 1995, Neural Computation.
[21] Henry S. Baird,et al. Document image defect models , 1995 .
[22] Christopher M. Bishop,et al. Current address: Microsoft Research, , 2022 .
[23] Bernhard Schölkopf,et al. Incorporating Invariances in Support Vector Learning Machines , 1996, ICANN.
[24] David J. Field,et al. Emergence of simple-cell receptive field properties by learning a sparse code for natural images , 1996, Nature.
[25] Guozhong An,et al. The Effects of Adding Noise During Backpropagation Training on a Generalization Performance , 1996, Neural Computation.
[26] Yves Grandvalet,et al. Noise Injection: Theoretical Prospects , 1997, Neural Computation.
[27] David J. Field,et al. Sparse coding with an overcomplete basis set: A strategy employed by V1? , 1997, Vision Research.
[28] H. Sebastian Seung,et al. Learning Continuous Attractors in Recurrent Networks , 1997, NIPS.
[29] Terrence J. Sejnowski,et al. The “independent components” of natural scenes are edge filters , 1997, Vision Research.
[30] David Maxwell Chickering,et al. Dependency Networks for Inference, Collaborative Filtering, and Data Visualization , 2000, J. Mach. Learn. Res..
[31] Nathalie Japkowicz,et al. Nonlinear Autoassociation Is Not Equivalent to PCA , 2000, Neural Computation.
[32] Paul E. Utgoff,et al. Many-Layered Learning , 2002, Neural Computation.
[33] Geoffrey E. Hinton. Training Products of Experts by Minimizing Contrastive Divergence , 2002, Neural Computation.
[34] H. Bourlard,et al. Auto-association by multilayer perceptrons and singular value decomposition , 1988, Biological Cybernetics.
[35] Johan Håstad,et al. On the power of small-depth threshold circuits , 1991, computational complexity.
[36] J. Bergstra. Algorithms for Classifying Recorded Music by Genre , 2006 .
[37] Yoshua Bengio,et al. Greedy Layer-Wise Training of Deep Networks , 2006, NIPS.
[38] Geoffrey E. Hinton,et al. Reducing the Dimensionality of Data with Neural Networks , 2006, Science.
[39] Yee Whye Teh,et al. A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.
[40] Marc'Aurelio Ranzato,et al. Efficient Learning of Sparse Representations with an Energy-Based Model , 2006, NIPS.
[41] Honglak Lee,et al. Sparse deep belief net model for visual area V2 , 2007, NIPS.
[42] Marc'Aurelio Ranzato,et al. Sparse Feature Learning for Deep Belief Networks , 2007, NIPS.
[43] Jason Weston,et al. Large-scale kernel machines , 2007 .
[44] Yoshua Bengio,et al. Scaling learning algorithms towards AI , 2007 .
[45] Yoshua Bengio,et al. An empirical evaluation of deep architectures on problems with many factors of variation , 2007, ICML '07.
[46] Yoshua. Bengio,et al. Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..
[47] Jason Weston,et al. Deep learning via semi-supervised embedding , 2008, ICML '08.
[48] Yoshua Bengio,et al. Extracting and composing robust features with denoising autoencoders , 2008, ICML '08.
[49] H. Sebastian Seung,et al. Natural Image Denoising with Convolutional Networks , 2008, NIPS.
[50] Yoshua Bengio,et al. Exploring Strategies for Training Deep Neural Networks , 2009, J. Mach. Learn. Res..
[51] Pascal Vincent,et al. Deep Learning using Robust Interdependent Codes , 2009, AISTATS.
[52] Lawrence K. Saul,et al. Kernel Methods for Deep Learning , 2009, NIPS.
[53] Yoshua Bengio,et al. Justifying and Generalizing Contrastive Divergence , 2009, Neural Computation.
[54] Yoshua Bengio,et al. Why Does Unsupervised Pre-training Help Deep Learning? , 2010, AISTATS.