Local Maxima in the Likelihood of Gaussian Mixture Models: Structural Results and Algorithmic Consequences
暂无分享,去创建一个
Martin J. Wainwright | Yuchen Zhang | Sivaraman Balakrishnan | Michael I. Jordan | Chi Jin | M. Wainwright | Chi Jin | Sivaraman Balakrishnan | Yuchen Zhang
[1] S. Yakowitz,et al. On the Identifiability of Finite Mixtures , 1968 .
[2] D. Rubin,et al. Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .
[3] A. F. Smith,et al. Statistical analysis of finite mixture distributions , 1986 .
[4] D. Donoho,et al. The "Automatic" Robustness of Minimum Distance Functionals , 1988 .
[5] Jiahua Chen. Optimal Rate of Convergence for Finite Mixture Models , 1995 .
[6] L. Wasserman,et al. RATES OF CONVERGENCE FOR THE GAUSSIAN MIXTURE SIEVE , 2000 .
[7] A. V. D. Vaart,et al. Entropies and rates of convergence for maximum likelihood and Bayes estimation for mixtures of normal densities , 2001 .
[8] Santosh S. Vempala,et al. A spectral algorithm for learning mixtures of distributions , 2002, The 43rd Annual IEEE Symposium on Foundations of Computer Science, 2002. Proceedings..
[9] Sanjeev Arora,et al. LEARNING MIXTURES OF SEPARATED NONSPHERICAL GAUSSIANS , 2005, math/0503457.
[10] Sanjoy Dasgupta,et al. A Probabilistic Analysis of EM for Mixtures of Separated, Spherical Gaussians , 2007, J. Mach. Learn. Res..
[11] Nathan Srebro. Are There Local Maxima in the Infinite-Sample Likelihood of Gaussian Mixture Estimation? , 2007, COLT.
[12] Satish Rao,et al. Learning Mixtures of Product Distributions Using Correlations and Independence , 2008, COLT.
[13] C. Matias,et al. Identifiability of parameters in latent structure models with many observed variables , 2008, 0809.5032.
[14] Ankur Moitra,et al. Settling the Polynomial Learnability of Mixtures of Gaussians , 2010, 2010 IEEE 51st Annual Symposium on Foundations of Computer Science.
[15] Mikhail Belkin,et al. Polynomial Learning of Distribution Families , 2010, 2010 IEEE 51st Annual Symposium on Foundations of Computer Science.
[16] Po-Ling Loh,et al. Regularized M-estimators with nonconvexity: statistical and algorithmic theory for local optima , 2013, J. Mach. Learn. Res..
[17] Sham M. Kakade,et al. Learning mixtures of spherical gaussians: moment methods and spectral decompositions , 2012, ITCS '13.
[18] Zhaoran Wang,et al. OPTIMAL COMPUTATIONAL AND STATISTICAL RATES OF CONVERGENCE FOR SPARSE NONCONVEX LEARNING PROBLEMS. , 2013, Annals of statistics.
[19] Surya Ganguli,et al. On the saddle point problem for non-convex optimization , 2014, ArXiv.
[20] Martin J. Wainwright,et al. Statistical guarantees for the EM algorithm: From population to sample-based analysis , 2014, ArXiv.
[21] Nhat Ho,et al. Identifiability and optimal rates of convergence for parameters of multiple types in finite mixtures , 2015, 1501.02497.
[22] M. Drton,et al. Maximum Likelihood Estimates for Gaussian Mixtures Are Transcendental , 2015, MACIS.
[23] Furong Huang,et al. Escaping From Saddle Points - Online Stochastic Gradient for Tensor Decomposition , 2015, COLT.
[24] Georgios Piliouras,et al. Gradient Descent Converges to Minimizers: The Case of Non-Isolated Critical Points , 2016, ArXiv.
[25] Michael I. Jordan,et al. Gradient Descent Converges to Minimizers , 2016, ArXiv.
[26] Georgios Piliouras,et al. Gradient Descent Only Converges to Minimizers: Non-Isolated Critical Points and Invariant Regions , 2016, ITCS.