General Low-rank Matrix Optimization: Geometric Analysis and Sharper Bounds
暂无分享,去创建一个
[1] Michael I. Jordan,et al. Accelerated Gradient Descent Escapes Saddle Points Faster than Gradient Descent , 2017, COLT.
[2] Renato D. C. Monteiro,et al. A nonlinear programming algorithm for solving semidefinite programs via low-rank factorization , 2003, Math. Program..
[3] Zhihui Zhu,et al. The Global Optimization Geometry of Low-Rank Matrix Optimization , 2017, IEEE Transactions on Information Theory.
[4] Yuxin Chen,et al. Nonconvex Optimization Meets Low-Rank Matrix Factorization: An Overview , 2018, IEEE Transactions on Signal Processing.
[5] Ohad Shamir,et al. Large-Scale Convex Minimization with a Low-Rank Constraint , 2011, ICML.
[6] Michael I. Jordan,et al. How to Escape Saddle Points Efficiently , 2017, ICML.
[7] Yi Zheng,et al. No Spurious Local Minima in Nonconvex Low Rank Problems: A Unified Geometric Analysis , 2017, ICML.
[8] Yudong Chen,et al. Harnessing Structures in Big Data via Guaranteed Low-Rank Matrix Estimation: Recent Theory and Fast Algorithms via Convex and Nonconvex Optimization , 2018, IEEE Signal Processing Magazine.
[9] Maxim Sviridenko,et al. Sparse Convex Optimization via Adaptively Regularized Hard Thresholding , 2020, ICML.
[10] Yuxin Chen,et al. Implicit Regularization in Nonconvex Statistical Estimation: Gradient Descent Converges Linearly for Phase Retrieval and Matrix Completion , 2018, ICML.
[11] Javad Lavaei,et al. On the Absence of Spurious Local Minima in Nonlinear Low-Rank Matrix Recovery Problems , 2020, AISTATS.
[12] Javad Lavaei,et al. How Much Restricted Isometry is Needed In Nonconvex Matrix Recovery? , 2018, NeurIPS.
[13] Pablo A. Parrilo,et al. Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization , 2007, SIAM Rev..
[14] Xiao Zhang,et al. A Unified Computational and Statistical Framework for Nonconvex Low-rank Matrix Estimation , 2016, AISTATS.
[15] Nicolas Boumal,et al. Nonconvex Phase Synchronization , 2016, SIAM J. Optim..
[16] Max Simchowitz,et al. Low-rank Solutions of Linear Matrix Equations via Procrustes Flow , 2015, ICML.
[17] Tengyu Ma,et al. Matrix Completion has No Spurious Local Minimum , 2016, NIPS.
[18] Nathan Srebro,et al. Global Optimality of Local Search for Low Rank Matrix Recovery , 2016, NIPS.
[19] Ji Chen,et al. Nonconvex Rectangular Matrix Completion via Gradient Descent Without ℓ₂,∞ Regularization , 2020, IEEE Transactions on Information Theory.
[20] John D. Lafferty,et al. A Convergent Gradient Descent Algorithm for Rank Minimization and Semidefinite Programming from Random Linear Measurements , 2015, NIPS.
[21] Yonina C. Eldar,et al. Phase Retrieval with Application to Optical Imaging: A contemporary overview , 2015, IEEE Signal Processing Magazine.
[22] A. Singer. Angular Synchronization by Eigenvectors and Semidefinite Programming. , 2009, Applied and computational harmonic analysis.
[23] John Wright,et al. A Geometric Analysis of Phase Retrieval , 2016, 2016 IEEE International Symposium on Information Theory (ISIT).
[24] Ewout van den Berg,et al. 1-Bit Matrix Completion , 2012, ArXiv.
[25] Inderjit S. Dhillon,et al. Guaranteed Rank Minimization via Singular Value Projection , 2009, NIPS.
[26] Georgios Piliouras,et al. Gradient Descent Only Converges to Minimizers: Non-Isolated Critical Points and Invariant Regions , 2016, ITCS.
[27] Emmanuel J. Candès,et al. Exact Matrix Completion via Convex Optimization , 2008, Found. Comput. Math..
[28] Javad Lavaei,et al. Sharp Restricted Isometry Bounds for the Inexistence of Spurious Local Minima in Nonconvex Matrix Recovery , 2019, J. Mach. Learn. Res..
[29] Benar Fux Svaiter,et al. Convergence of descent methods for semi-algebraic and tame problems: proximal algorithms, forward–backward splitting, and regularized Gauss–Seidel methods , 2013, Math. Program..
[30] Emmanuel J. Candès,et al. The Power of Convex Relaxation: Near-Optimal Matrix Completion , 2009, IEEE Transactions on Information Theory.
[31] Qiuwei Li,et al. The non-convex geometry of low-rank matrix optimization , 2016, Information and Inference: A Journal of the IMA.
[32] Michael I. Jordan,et al. Gradient Descent Only Converges to Minimizers , 2016, COLT.
[33] Haoyang Liu,et al. An equivalence between stationary points for rank constraints versus low-rank factorizations , 2018, 1812.00404.
[34] Zhihui Zhu,et al. Global Optimality in Low-Rank Matrix Optimization , 2017, IEEE Transactions on Signal Processing.
[35] Stephen Becker,et al. Perturbed Proximal Descent to Escape Saddle Points for Non-convex and Non-smooth Objective Functions , 2019, INNSBDDL.
[36] Anastasios Kyrillidis,et al. Dropping Convexity for Faster Semi-definite Optimization , 2015, COLT.