Deep belief nets are probabilistic generative models that are composed of multiple layers of stochastic latent variables (also called “feature detectors” or “hidden units”). The top two layers have undirected, symmetric connections between them and form an associative memory. The lower layers receive top-down, directed connections from the layer above. Deep belief nets have two important computational properties. First, there is an efficient procedure for learning the topdown, generative weights that specify how the variables in one layer determine the probabilities of variables in the layer below. This procedure learns one layer of latent variables at a time. Second, after learning multiple layers, the values of the latent variables in every layer can be inferred by a single, bottom-up pass that starts with an observed data vector in the bottom layer and uses the generative weights in the reverse direction.
[1]
Geoffrey E. Hinton.
Connectionist Learning Procedures
,
1989,
Artif. Intell..
[2]
Yee Whye Teh,et al.
A Fast Learning Algorithm for Deep Belief Nets
,
2006,
Neural Computation.
[3]
Yoshua Bengio,et al.
An empirical evaluation of deep architectures on problems with many factors of variation
,
2007,
ICML '07.
[4]
Marc'Aurelio Ranzato,et al.
Unsupervised Learning of Invariant Feature Hierarchies with Applications to Object Recognition
,
2007,
2007 IEEE Conference on Computer Vision and Pattern Recognition.
[5]
Geoffrey E. Hinton,et al.
Learning representations by back-propagating errors
,
1986,
Nature.
[6]
Geoffrey E. Hinton,et al.
Reducing the Dimensionality of Data with Neural Networks
,
2006,
Science.