A novel variational form of the Schatten-p quasi-norm

The Schatten-$p$ quasi-norm with $p\in(0,1)$ has recently gained considerable attention in various low-rank matrix estimation problems offering significant benefits over relevant convex heuristics such as the nuclear norm. However, due to the nonconvexity of the Schatten-$p$ quasi-norm, minimization suffers from two major drawbacks: 1) the lack of theoretical guarantees and 2) the high computational cost which is demanded for the minimization task even for trivial tasks such as finding stationary points. In an attempt to reduce the high computational cost induced by Schatten-$p$ quasi-norm minimization, variational forms, which are defined over smaller-size matrix factors whose product equals the original matrix, have been proposed. Here, we propose and analyze a novel variational form of Schatten-$p$ quasi-norm which, for the first time in the literature, is defined for any continuous value of $p\in(0,1]$ and decouples along the columns of the factorized matrices. The proposed form can be considered as the natural generalization of the well-known variational form of the nuclear norm to the nonconvex case i.e., for $p\in(0,1)$. The resulting formulation gives way to SVD-free algorithms thus offering lower computational complexity than the one that is induced by the original definition of the Schatten-$p$ quasi-norm. A local optimality analysis is provided which shows~that we can arrive at a local minimum of the original Schatten-$p$ quasi-norm problem by reaching a local minimum of the matrix factorization based surrogate problem. In addition, for the case of the squared Frobenius loss with linear operators obeying the restricted isometry property (RIP), a rank-one update scheme is proposed, which offers a way to escape poor local minima. Finally, the efficiency of our approach is empirically shown on a matrix completion problem.

[1]  Shuicheng Yan,et al.  Nonconvex Nonsmooth Low Rank Minimization via Iteratively Reweighted Nuclear Norm , 2015, IEEE Transactions on Image Processing.

[2]  Yan Liu,et al.  Weighted Schatten $p$ -Norm Minimization for Image Denoising and Background Subtraction , 2015, IEEE Transactions on Image Processing.

[3]  Zhi-Quan Luo,et al.  A Unified Algorithmic Framework for Block-Structured Optimization Involving Big Data: With applications in machine learning and signal processing , 2015, IEEE Signal Processing Magazine.

[4]  Mitsuru Uchiyama,et al.  Subadditivity of eigenvalue sums , 2005 .

[5]  Trevor J. Hastie,et al.  Matrix completion and low-rank SVD via fast alternating least squares , 2014, J. Mach. Learn. Res..

[6]  Feiping Nie,et al.  Low-Rank Matrix Recovery via Efficient Schatten p-Norm Minimization , 2012, AAAI.

[7]  Nathan Srebro,et al.  Fast maximum margin matrix factorization for collaborative prediction , 2005, ICML.

[8]  Alexandre Bernardino,et al.  Unifying Nuclear Norm and Bilinear Factorization Approaches for Low-Rank Matrix Decomposition , 2013, 2013 IEEE International Conference on Computer Vision.

[9]  Simon Foucart,et al.  Concave Mirsky Inequality and Low-Rank Recovery , 2018, SIAM J. Matrix Anal. Appl..

[10]  Yuanyuan Liu,et al.  Scalable Algorithms for Tractable Schatten Quasi-Norm Minimization , 2016, AAAI.

[11]  Zhi-Quan Luo,et al.  A Unified Convergence Analysis of Block Successive Minimization Methods for Nonsmooth Optimization , 2012, SIAM J. Optim..

[12]  Lu Liu,et al.  Exact minimum rank approximation via Schatten p-norm minimization , 2014, J. Comput. Appl. Math..

[13]  René Vidal,et al.  Global Optimality in Neural Network Training , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[14]  René Vidal,et al.  Structured Low-Rank Matrix Factorization: Global Optimality, Algorithms, and Applications , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[15]  Yuanyuan Liu,et al.  Tractable and Scalable Schatten Quasi-Norm Approximations for Rank Minimization , 2016, AISTATS.

[16]  Evan Schwab,et al.  Global Optimality in Separable Dictionary Learning with Applications to the Analysis of Diffusion MRI , 2019, SIAM J. Imaging Sci..

[17]  R. C. Thompson Convex and concave functions of singular values of matrix sums. , 1976 .

[18]  Renato D. C. Monteiro,et al.  Digital Object Identifier (DOI) 10.1007/s10107-004-0564-1 , 2004 .

[19]  Hongbin Zha,et al.  A Unified Convex Surrogate for the Schatten-p Norm , 2016, AAAI.

[20]  Jicong Fan,et al.  Factor Group-Sparse Regularization for Efficient Low-Rank Matrix Recovery , 2019, NeurIPS.

[21]  Hristo S. Sendov,et al.  Nonsmooth Analysis of Singular Values. Part I: Theory , 2005 .

[22]  Yuanyuan Liu,et al.  Unified Scalable Equivalent Formulations for Schatten Quasi-Norms , 2016, ArXiv.

[23]  Jian Yang,et al.  LRR for Subspace Segmentation via Tractable Schatten- $p$ Norm Minimization and Factorization , 2019, IEEE Transactions on Cybernetics.

[24]  Anders Heyden,et al.  Bilinear Parameterization For Differentiable Rank-Regularization , 2018, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[25]  Paris V. Giampouras,et al.  Alternating Iteratively Reweighted Least Squares Minimization for Low-Rank Matrix Factorization , 2019, IEEE Transactions on Signal Processing.