Provably Learning Mixtures of Gaussians and More

Given a random sample, how can one accurately estimate the parameters of the probabilistic model that generated the data? This is a fundamental question in statistical inference and, more recently, in machine learning. One of the most widely studied instances of this problem is estimating the parameters of a mixture of Gaussians, since doing so is of fundamental importance in a wide range of subjects, from physics to social science. The classic, and most popular approach, for learning Gaussian mixture models (GMMs) is the EM algorithm [Dempster et al., 1977]. The EM algorithm is a search heuristic over the parameters that finds a local maximum of the likelihood function, and therefore makes no guarantees that it will converge to an estimate that is close to the true parameters. Furthermore, in practice it has been found to converge very slowly.

[1]  Sanjeev Arora,et al.  Learning mixtures of arbitrary gaussians , 2001, STOC '01.

[2]  Sanjoy Dasgupta,et al.  Learning mixtures of Gaussians , 1999, 40th Annual Symposium on Foundations of Computer Science (Cat. No.99CB37039).

[3]  Santosh S. Vempala,et al.  A spectral algorithm for learning mixture models , 2004, J. Comput. Syst. Sci..

[4]  Ankur Moitra,et al.  Settling the Polynomial Learnability of Mixtures of Gaussians , 2010, 2010 IEEE 51st Annual Symposium on Foundations of Computer Science.

[5]  Dimitris Achlioptas,et al.  On Spectral Learning of Mixtures of Distributions , 2005, COLT.

[6]  Satish Rao,et al.  Learning Mixtures of Product Distributions Using Correlations and Independence , 2008, COLT.

[7]  Satish Rao,et al.  Beyond Gaussians: Spectral Methods for Learning Mixtures of Heavy-Tailed Distributions , 2008, COLT.

[8]  Mikhail Belkin,et al.  Toward Learning Gaussian Mixtures with Arbitrary Separation , 2010, COLT.

[9]  Jon M. Kleinberg,et al.  On learning mixtures of heavy-tailed distributions , 2005, 46th Annual IEEE Symposium on Foundations of Computer Science (FOCS'05).

[10]  J. Feldman,et al.  Learning mixtures of product distributions over discrete domains , 2005, 46th Annual IEEE Symposium on Foundations of Computer Science (FOCS'05).

[11]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .

[12]  Sanjoy Dasgupta,et al.  A Two-Round Variant of EM for Gaussian Mixtures , 2000, UAI.

[13]  Mikhail Belkin,et al.  Polynomial Learning of Distribution Families , 2010, 2010 IEEE 51st Annual Symposium on Foundations of Computer Science.

[14]  Jon Feldman,et al.  PAC Learning Axis-Aligned Mixtures of Gaussians with No Separation Assumption , 2006, COLT.

[15]  Santosh S. Vempala,et al.  The Spectral Method for General Mixture Models , 2005, COLT.

[16]  Adam Tauman Kalai,et al.  Efficiently learning mixtures of two Gaussians , 2010, STOC '10.