Where Do Features Come From?
暂无分享,去创建一个
[1] Robert B. Ash,et al. Information Theory , 2020, The SAGE International Encyclopedia of Mass Media and Society.
[2] L. Baum,et al. An inequality and associated maximization technique in statistical estimation of probabilistic functions of a Markov process , 1972 .
[3] P. Werbos,et al. Beyond Regression : "New Tools for Prediction and Analysis in the Behavioral Sciences , 1974 .
[4] Berthold K. P. Horn. Understanding Image Intensities , 1977, Artif. Intell..
[5] D. Rubin,et al. Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .
[6] James L. McClelland,et al. An interactive activation model of context effects in letter perception: I. An account of basic findings. , 1981 .
[7] Francis Crick,et al. The function of dream sleep , 1983, Nature.
[8] David Heckerman,et al. Probabilistic Interpretation for MYCIN's Certainty Factors , 1990, UAI.
[9] Yann LeCun,et al. Une procedure d'apprentissage pour reseau a seuil asymmetrique (A learning scheme for asymmetric threshold networks) , 1985 .
[10] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[11] Paul Smolensky,et al. Information processing in dynamical systems: foundations of harmony theory , 1986 .
[12] Geoffrey E. Hinton,et al. Learning and relearning in Boltzmann machines , 1986 .
[13] James L. McClelland,et al. Parallel distributed processing: explorations in the microstructure of cognition, vol. 1: foundations , 1986 .
[14] David J. Spiegelhalter,et al. Local computations with probabilities on graphical structures and their application to expert systems , 1990 .
[15] James L. McClelland,et al. An interactive activation model of context effects in letter perception: part 1.: an account of basic findings , 1988 .
[16] Judea Pearl,et al. Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.
[17] Geoffrey E. Hinton. Connectionist Learning Procedures , 1989, Artif. Intell..
[18] Geoffrey E. Hinton,et al. Phoneme recognition using time-delay neural networks , 1989, IEEE Trans. Acoust. Speech Signal Process..
[19] Geoffrey E. Hinton,et al. Distributed Representations , 1986, The Philosophy of Artificial Intelligence.
[20] Jeffrey L. Elman,et al. Finding Structure in Time , 1990, Cogn. Sci..
[21] Radford M. Neal. Connectionist Learning of Belief Networks , 1992, Artif. Intell..
[22] Geoffrey E. Hinton,et al. Autoencoders, Minimum Description Length and Helmholtz Free Energy , 1993, NIPS.
[23] J. O’Keefe,et al. Phase relationship between hippocampal place units and the EEG theta rhythm , 1993, Hippocampus.
[24] Yoshua Bengio,et al. Learning long-term dependencies with gradient descent is difficult , 1994, IEEE Trans. Neural Networks.
[25] Geoffrey E. Hinton,et al. The "wake-sleep" algorithm for unsupervised neural networks. , 1995, Science.
[26] R Hecht-Nielsen,et al. Replicator neural networks for universal optimal source coding. , 1995, Science.
[27] Geoffrey E. Hinton,et al. Bayesian Learning for Neural Networks , 1995 .
[28] Michael I. Jordan,et al. Mean Field Theory for Sigmoid Belief Networks , 1996, J. Artif. Intell. Res..
[29] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[30] H. Markram,et al. Regulation of Synaptic Efficacy by Coincidence of Postsynaptic APs and EPSPs , 1997, Science.
[31] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[32] Geoffrey E. Hinton,et al. A View of the Em Algorithm that Justifies Incremental, Sparse, and other Variants , 1998, Learning in Graphical Models.
[33] Michael I. Jordan,et al. An Introduction to Variational Methods for Graphical Models , 1999, Machine-mediated learning.
[34] X. Yao. Evolving Artificial Neural Networks , 1999 .
[35] Geoffrey E. Hinton. Training Products of Experts by Minimizing Contrastive Divergence , 2002, Neural Computation.
[36] Geoffrey E. Hinton,et al. Exponential Family Harmoniums with an Application to Information Retrieval , 2004, NIPS.
[37] David J. C. MacKay,et al. Information Theory, Inference, and Learning Algorithms , 2004, IEEE Transactions on Information Theory.
[38] J. Tenenbaum,et al. Theory-based Bayesian models of inductive learning and reasoning , 2006, Trends in Cognitive Sciences.
[39] Geoffrey E. Hinton,et al. Reducing the Dimensionality of Data with Neural Networks , 2006, Science.
[40] Karl J. Friston,et al. A free energy principle for the brain , 2006, Journal of Physiology-Paris.
[41] Yee Whye Teh,et al. A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.
[42] Marc'Aurelio Ranzato,et al. Sparse Feature Learning for Deep Belief Networks , 2007, NIPS.
[43] Michel Verleysen,et al. Nonlinear Dimensionality Reduction , 2021, Computer Vision.
[44] Tijmen Tieleman,et al. Training restricted Boltzmann machines using approximations to the likelihood gradient , 2008, ICML '08.
[45] Geoffrey E. Hinton,et al. Using fast weights to improve persistent contrastive divergence , 2009, ICML '09.
[46] Honglak Lee,et al. Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations , 2009, ICML '09.
[47] Carl E. Rasmussen,et al. Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.
[48] Yoshua Bengio,et al. Why Does Unsupervised Pre-training Help Deep Learning? , 2010, AISTATS.
[49] James Martens,et al. Deep learning via Hessian-free optimization , 2010, ICML.
[50] Geoffrey E. Hinton,et al. Phone Recognition with the Mean-Covariance Restricted Boltzmann Machine , 2010, NIPS.
[51] Geoffrey E. Hinton,et al. Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.
[52] Pascal Vincent,et al. Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion , 2010, J. Mach. Learn. Res..
[53] Peggy Seriès,et al. Hallucinations in Charles Bonnet Syndrome Induced by Homeostasis: a Deep Boltzmann Machine Model , 2010, NIPS.
[54] Geoffrey E. Hinton. Learning to represent visual input , 2010, Philosophical Transactions of the Royal Society B: Biological Sciences.
[55] Pascal Vincent,et al. Contractive Auto-Encoders: Explicit Invariance During Feature Extraction , 2011, ICML.
[56] Geoffrey E. Hinton,et al. Transforming Auto-Encoders , 2011, ICANN.
[57] Geoffrey E. Hinton,et al. Two Distributed-State Models For Generating High-Dimensional Time Series , 2011, J. Mach. Learn. Res..
[58] Wolfgang Maass,et al. Neural Dynamics as Sampling: A Model for Stochastic Computation in Recurrent Networks of Spiking Neurons , 2011, PLoS Comput. Biol..
[59] Geoffrey E. Hinton,et al. An Efficient Learning Procedure for Deep Boltzmann Machines , 2012, Neural Computation.