The generative aspect model is an extension of the multinomial model for text that allows word probabilities to vary stochastically across documents. Previous results with aspect models have been promising, but hindered by the computational difficulty of carrying out inference and learning. This paper demonstrates that the simple variational methods of Blei et al. (2001) can lead to inaccurate inferences and biased learning for the generative aspect model. We develop an alternative approach that leads to higher accuracy at comparable cost. An extension of Expectation-Propagation is used for inference and then embedded in an EM algorithm for learning. Experimental results are presented for both synthetic and real data sets.
[1]
A. Dawid,et al.
A comparison of sequential learning methods for incomplete data
,
1995
.
[2]
Thomas Hofmann,et al.
Probabilistic Latent Semantic Analysis
,
1999,
UAI.
[3]
Hagai Attias,et al.
A Variational Bayesian Framework for Graphical Models
,
1999
.
[4]
Thomas P. Minka,et al.
Using lower bounds to approxi-mate integrals
,
2001
.
[5]
Tom Minka,et al.
A family of algorithms for approximate Bayesian inference
,
2001
.
[6]
M. Curr.
In our case
,
2001
.
[7]
Michael I. Jordan,et al.
Latent Dirichlet Allocation
,
2001,
J. Mach. Learn. Res..