Double robust principal component analysis

Abstract Robust Principal Component Analysis (RPCA) aiming to recover underlying clean data with low-rank structure from the corrupted data, is a powerful tool in machine learning and data mining. However, in many real-world applications where new data (i.e., out-of-samples) in the testing phase can be unseen in the training procedure, (1) RPCA which is a transductive method can be naturally incapable of handing out-of-samples, and (2) violently applying RPCA into this applications does not explicitly consider the relationships between reconstruction error and low-rank representation. To tackle these problems, in this paper, we propose a Double Robust Principal Component Analysis to deal with the out-of-sample problems, which is termed as DRPCA. More specifically, we integrate a reconstruction error into the criterion function of RPCA. Our proposed model can then benefit from (1) the robustness of principal components to outliers and missing values, (2) the bridge between reconstruction error and low-rank representation, (3) low-rank clean data extraction from new datum by a linear transform. To this end, extensive experiments on several datasets demonstrate its superiority, when comparing with the state-of-the-art models, in several clustering and low-rank recovery tasks.

[1]  Rui Liu,et al.  Density-sensitive Robust Fuzzy Kernel Principal Component Analysis technique , 2019, Neurocomputing.

[2]  Lucia Maddalena,et al.  Scene background initialization: A taxonomy , 2017, Pattern Recognit. Lett..

[3]  David J. Kriegman,et al.  Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection , 1996, ECCV.

[4]  B Anton,et al.  Immunohistochemical localization of ORL‐1 in the central nervous system of the rat , 1996, The Journal of comparative neurology.

[5]  Yinglin Wang,et al.  Low rank approximation with sparse integration of multiple manifolds for data representation , 2014, Applied Intelligence.

[6]  Wei Liu,et al.  Tensor Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Tensors via Convex Optimization , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Feiping Nie,et al.  Angle 2DPCA: A New Formulation for 2DPCA , 2018, IEEE Transactions on Cybernetics.

[8]  Bernhard Schölkopf,et al.  Nonlinear Component Analysis as a Kernel Eigenvalue Problem , 1998, Neural Computation.

[9]  Fang Liu,et al.  Global Low-Rank Image Restoration With Gaussian Mixture Model , 2018, IEEE Transactions on Cybernetics.

[10]  Rong Wang,et al.  Robust 2DPCA With Non-greedy $\ell _{1}$ -Norm Maximization for Image Analysis , 2015, IEEE Transactions on Cybernetics.

[11]  Jing Wang,et al.  Trace ratio 2DLDA with L1-norm optimization , 2017, Neurocomputing.

[12]  Lucia Maddalena,et al.  Towards Benchmarking Scene Background Initialization , 2015, ICIAP Workshops.

[13]  Takeo Kanade,et al.  Robust L/sub 1/ norm factorization in the presence of outliers and missing data by alternative convex programming , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[14]  Qianqian Wang,et al.  Two-Dimensional PCA with F-Norm Minimization , 2017, AAAI.

[15]  Xiaoming Yuan,et al.  Sparse and low-rank matrix decomposition via alternating direction method , 2013 .

[16]  Jane You,et al.  Low-rank matrix factorization with multiple Hypergraph regularizer , 2015, Pattern Recognit..

[17]  In-So Kweon,et al.  Partial Sum Minimization of Singular Values in Robust PCA: Algorithm and Applications , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[18]  Nojun Kwak,et al.  Principal Component Analysis Based on L1-Norm Maximization , 2008, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[19]  Xavier Bresson,et al.  Robust Principal Component Analysis on Graphs , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[20]  Meng Wang,et al.  Constrained Low-Rank Learning Using Least Squares-Based Regularization , 2016, IEEE Transactions on Cybernetics.

[21]  Jin Tang,et al.  Graph-Laplacian PCA: Closed-Form Solution and Robustness , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[22]  Feiping Nie,et al.  On the schatten norm for matrix based subspace learning and classification , 2016, Neurocomputing.

[23]  Zhenyue Zhang,et al.  Low-Rank Matrix Approximation with Manifold Regularization , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[24]  Stephen P. Boyd,et al.  Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers , 2011, Found. Trends Mach. Learn..

[25]  Yi Ma,et al.  Robust principal component analysis? , 2009, JACM.

[26]  Chris H. Q. Ding,et al.  R1-PCA: rotational invariant L1-norm principal component analysis for robust subspace factorization , 2006, ICML.

[27]  Shuicheng Yan,et al.  Neighborhood preserving embedding , 2005, Tenth IEEE International Conference on Computer Vision (ICCV'05) Volume 1.

[28]  Feiping Nie,et al.  $\ell _{2,p}$ -Norm Based PCA for Image Recognition , 2018, IEEE Transactions on Image Processing.

[29]  M. Turk,et al.  Eigenfaces for Recognition , 1991, Journal of Cognitive Neuroscience.

[30]  Ran He,et al.  Robust Principal Component Analysis Based on Maximum Correntropy Criterion , 2011, IEEE Transactions on Image Processing.

[31]  Emmanuel J. Candès,et al.  A Singular Value Thresholding Algorithm for Matrix Completion , 2008, SIAM J. Optim..

[32]  Xiaofei He,et al.  Locality Preserving Projections , 2003, NIPS.

[33]  Yang Liu,et al.  Learning more distinctive representation by enhanced PCA network , 2018, Neurocomputing.