Low-Rank Matrix Recovery via Continuation-Based Approximate Low-Rank Minimization

Low-rank matrix recovery (LRMR) is to recover the underlying low-rank structure from the degraded observations, and has myriad formulations, for example robust principal component analysis (RPCA). As a core component of LRMR, the low-rank approximation model attempts to capture the low-rank structure by approximating the \(\ell _0\)-norm of all the singular values of a low-rank matrix, i.e., the number of the non-zero singular values. Towards this purpose, this paper develops a low-rank approximation model by jointly combining a parameterized hyperbolic tangent (tanh) function with the continuation process. Specificially, the continuation process is exploited to impose the parameterized tanh function to approximate the \(\ell _0\)-norm. We then apply the proposed low-rank model to RPCA and refer to it as tanh-RPCA. Convergence analysis on optimization, and experiments of background subtraction on seven challenging real-world videos show the efficacy of the proposed low-rank model through comparing tanh-RPCA with several state-of-the-art methods.

[1]  Qi Tian,et al.  Statistical modeling of complex backgrounds for foreground object detection , 2004, IEEE Transactions on Image Processing.

[2]  Xiaojie Guo,et al.  ROUTE: Robust Outlier Estimation for Low Rank Matrix Recovery , 2017, IJCAI.

[3]  Zhao Kang,et al.  Robust PCA Via Nonconvex Rank Approximation , 2015, 2015 IEEE International Conference on Data Mining.

[4]  Jingdong Wang,et al.  A Probabilistic Approach to Robust Matrix Factorization , 2012, ECCV.

[5]  Ming Yang,et al.  RAP: Scalable RPCA for Low-rank Matrix Recovery , 2016, CIKM.

[6]  Jieping Ye,et al.  Multi-stage multi-task feature learning , 2012, J. Mach. Learn. Res..

[7]  Yu-Chiang Frank Wang,et al.  Low-rank matrix recovery with structural incoherence for robust face recognition , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[8]  Zhi Liu,et al.  Segmentation Driven Low-rank Matrix Recovery for Saliency Detection , 2013, BMVC.

[9]  Jian Dong,et al.  Accelerated low-rank visual recovery by random projection , 2011, CVPR 2011.

[10]  Yuling Jiao,et al.  A Primal Dual Active Set Algorithm With Continuation for Compressed Sensing , 2013, IEEE Transactions on Signal Processing.

[11]  Zhao Kang,et al.  Subspace Clustering Using Log-determinant Rank Approximation , 2015, KDD.

[12]  Pablo A. Parrilo,et al.  Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization , 2007, SIAM Rev..

[13]  Guillermo Sapiro,et al.  Online Learning for Matrix Factorization and Sparse Coding , 2009, J. Mach. Learn. Res..

[14]  Cun-Hui Zhang Nearly unbiased variable selection under minimax concave penalty , 2010, 1002.4734.

[15]  Ying Wu,et al.  A unified approach to salient object detection via low rank matrix recovery , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[16]  Marc Teboulle,et al.  A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems , 2009, SIAM J. Imaging Sci..

[17]  Long Lan,et al.  Background Subtraction via 3D Convolutional Neural Networks , 2018, 2018 24th International Conference on Pattern Recognition (ICPR).

[18]  Zhao Kang,et al.  Robust Subspace Clustering via Smoothed Rank Approximation , 2015, IEEE Signal Processing Letters.

[19]  Arvind Ganesh,et al.  Fast Convex Optimization Algorithms for Exact Recovery of a Corrupted Low-Rank Matrix , 2009 .

[20]  Lei Zhang,et al.  Weighted Nuclear Norm Minimization and Its Applications to Low Level Vision , 2016, International Journal of Computer Vision.

[21]  Zhixun Su,et al.  Linearized Alternating Direction Method with Adaptive Penalty for Low-Rank Representation , 2011, NIPS.

[22]  Shuicheng Yan,et al.  Generalized Nonconvex Nonsmooth Low-Rank Minimization , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[23]  Shuicheng Yan,et al.  Practical low-rank matrix approximation under robust L1-norm , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[24]  Kentaro Toyama,et al.  Wallflower: principles and practice of background maintenance , 1999, Proceedings of the Seventh IEEE International Conference on Computer Vision.

[25]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[26]  Hristo S. Sendov,et al.  Nonsmooth Analysis of Singular Values. Part I: Theory , 2005 .

[27]  Yi Ma,et al.  Robust principal component analysis? , 2009, JACM.

[28]  Emmanuel J. Candès,et al.  Exact Matrix Completion via Convex Optimization , 2009, Found. Comput. Math..

[29]  Li Zhang,et al.  Robust Alternating Low-Rank Representation by joint Lp- and L2, p-norm minimization , 2017, Neural Networks.

[30]  Jieping Ye,et al.  Efficient Sparse Group Feature Selection via Nonconvex Optimization , 2012, ICML.

[31]  Xiaoming Yuan,et al.  Sparse and low-rank matrix decomposition via alternating direction method , 2013 .

[32]  Tong Zhang,et al.  Analysis of Multi-stage Convex Relaxation for Sparse Regularization , 2010, J. Mach. Learn. Res..

[33]  Wen Gao,et al.  Background Subtraction via generalized fused lasso foreground modeling , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[34]  Philip S. Yu,et al.  HashNet: Deep Learning to Hash by Continuation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[35]  John Wright,et al.  Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Matrices via Convex Optimization , 2009, NIPS.

[36]  Shuicheng Yan,et al.  Similarity preserving low-rank representation for enhanced data representation and effective subspace learning , 2014, Neural Networks.

[37]  Zhao Kang,et al.  Robust Subspace Clustering via Tighter Rank Approximation , 2015, CIKM.

[38]  Xuelong Li,et al.  Fast and Accurate Matrix Completion via Truncated Nuclear Norm Regularization , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[39]  Zhongping Wan,et al.  An Alternating Direction Method with Continuation for Nonconvex Low Rank Minimization , 2016, J. Sci. Comput..