Diffusion Approximations for Online Principal Component Estimation and Global Convergence
暂无分享,去创建一个
[1] Ohad Shamir,et al. Fast Stochastic Algorithms for SVD and PCA: Convergence Properties and Convexity , 2015, ICML.
[2] Xiao-Tong Yuan,et al. Truncated power method for sparse eigenvalue problems , 2011, J. Mach. Learn. Res..
[3] Michael I. Jordan,et al. Gradient Descent Only Converges to Minimizers , 2016, COLT.
[4] B. Nadler. Finite sample approximation results for principal component analysis: a matrix perturbation approach , 2009, 0901.3245.
[5] John Wright,et al. Complete Dictionary Recovery Over the Sphere II: Recovery by Riemannian Trust-Region Method , 2015, IEEE Transactions on Information Theory.
[6] Christopher De Sa,et al. Global Convergence of Stochastic Gradient Descent for Some Non-convex Matrix Problems , 2014, ICML.
[7] Georgios Piliouras,et al. Gradient Descent Only Converges to Minimizers: Non-Isolated Critical Points and Invariant Regions , 2016, ITCS.
[8] U. Helmke,et al. Optimization and Dynamical Systems , 1994, Proceedings of the IEEE.
[9] Georgios Piliouras,et al. Gradient Descent Converges to Minimizers: The Case of Non-Isolated Critical Points , 2016, ArXiv.
[10] Elad Hazan,et al. Fast and Simple PCA via Convex Optimization , 2015, ArXiv.
[11] E Weinan,et al. Stochastic Modified Equations and Adaptive Stochastic Gradient Algorithms , 2015, ICML.
[12] M. Wainwright,et al. High-dimensional analysis of semidefinite relaxations for sparse principal components , 2008, 2008 IEEE International Symposium on Information Theory.
[13] J. Kuczy,et al. Estimating the Largest Eigenvalue by the Power and Lanczos Algorithms with a Random Start , 1992 .
[14] S. Shreve,et al. Stochastic differential equations , 1955, Mathematical Proceedings of the Cambridge Philosophical Society.
[15] Sanjoy Dasgupta,et al. The Fast Convergence of Incremental PCA , 2013, NIPS.
[16] E. Oja,et al. On stochastic approximation of the eigenvectors and eigenvalues of the expectation of a random matrix , 1985 .
[17] Cameron Musco,et al. Randomized Block Krylov Methods for Stronger and Faster Approximate Singular Value Decomposition , 2015, NIPS.
[18] John E. Moody,et al. Towards Faster Stochastic Gradient Search , 1991, NIPS.
[19] John Wright,et al. Complete Dictionary Recovery Over the Sphere I: Overview and the Geometric Picture , 2015, IEEE Transactions on Information Theory.
[20] Furong Huang,et al. Escaping From Saddle Points - Online Stochastic Gradient for Tensor Decomposition , 2015, COLT.
[21] Ohad Shamir,et al. A Stochastic PCA and SVD Algorithm with an Exponential Convergence Rate , 2014, ICML.
[22] John Wright,et al. A Geometric Analysis of Phase Retrieval , 2016, 2016 IEEE International Symposium on Information Theory (ISIT).
[23] Jing Lei,et al. Minimax Rates of Estimation for Sparse PCA in High Dimensions , 2012, AISTATS.
[24] John Wright,et al. When Are Nonconvex Problems Not Scary? , 2015, ArXiv.
[25] Ohad Shamir,et al. Convergence of Stochastic Gradient Descent for PCA , 2015, ICML.
[26] Alexandre d'Aspremont,et al. Optimal Solutions for Sparse Principal Component Analysis , 2007, J. Mach. Learn. Res..
[27] V. Climenhaga. Markov chains and mixing times , 2013 .
[28] Vincent Q. Vu,et al. MINIMAX SPARSE PRINCIPAL SUBSPACE ESTIMATION IN HIGH DIMENSIONS , 2012, 1211.0373.
[29] Zongming Ma. Sparse Principal Component Analysis and Iterative Thresholding , 2011, 1112.2432.
[30] B. Øksendal. Stochastic Differential Equations , 1985 .
[31] Tong Zhang,et al. Near-optimal stochastic approximation for online principal component estimation , 2016, Math. Program..
[32] Anima Anandkumar,et al. Efficient approaches for escaping higher order saddle points in non-convex optimization , 2016, COLT.
[33] I. Johnstone,et al. On Consistency and Sparsity for Principal Components Analysis in High Dimensions , 2009, Journal of the American Statistical Association.
[34] Nathan Srebro,et al. Stochastic optimization for PCA and PLS , 2012, 2012 50th Annual Allerton Conference on Communication, Control, and Computing (Allerton).
[35] Moritz Hardt,et al. The Noisy Power Method: A Meta Algorithm with Applications , 2013, NIPS.
[36] Nathan Srebro,et al. Stochastic Optimization of PCA with Capped MSG , 2013, NIPS.
[37] H. Zou. The Adaptive Lasso and Its Oracle Properties , 2006 .
[38] D. Aldous. Probability Approximations via the Poisson Clumping Heuristic , 1988 .
[39] Zhaoran Wang,et al. Nonconvex Statistical Optimization: Minimax-Optimal Sparse PCA in Polynomial Time , 2014, ArXiv.
[40] S. Ethier,et al. Markov Processes: Characterization and Convergence , 2005 .
[41] Cameron Musco,et al. Stronger Approximate Singular Value Decomposition via the Block Lanczos and Power Methods , 2015, ArXiv.
[42] E Weinan,et al. Dynamics of Stochastic Gradient Algorithms , 2015, ArXiv.
[43] Prateek Jain,et al. Streaming PCA: Matching Matrix Bernstein and Near-Optimal Finite Sample Guarantees for Oja's Algorithm , 2016, COLT.
[44] Ioannis Mitliagkas,et al. Memory Limited, Streaming PCA , 2013, NIPS.
[45] Stephen P. Boyd,et al. A Differential Equation for Modeling Nesterov's Accelerated Gradient Method: Theory and Insights , 2014, J. Mach. Learn. Res..
[46] T. P. Krasulina. The method of stochastic approximation for the determination of the least eigenvalue of a symmetrical matrix , 1969 .
[47] David M. Blei,et al. A Variational Analysis of Stochastic Gradient Algorithms , 2016, ICML.
[48] T. Cai,et al. Sparse PCA: Optimal rates and adaptive estimation , 2012, 1211.1309.
[49] E. Oja. Simplified neuron model as a principal component analyzer , 1982, Journal of mathematical biology.
[50] Prateek Jain,et al. Matching Matrix Bernstein with Little Memory: Near-Optimal Finite Sample Guarantees for Oja's Algorithm , 2016, ArXiv.