A Trimmed Clustering-Based l1-Principal Component Analysis Model for Image Classification and Clustering Problems with Outliers

Different versions of principal component analysis (PCA) have been widely used to extract important information for image recognition and image clustering problems. However, owing to the presence of outliers, this remains challenging. This paper proposes a new PCA methodology based on a novel discovery that the widely used l1-PCA is equivalent to a two-groups k-means clustering model. The projection vector of the l1-PCA is the vector difference between the two cluster centers estimated by the clustering model. In theory, this vector difference provides inter-cluster information, which is beneficial for distinguishing data objects from different classes. However, the performance of l1-PCA is not comparable with the state-of-the-art methods. This is because the l1-PCA can be sensitive to outliers, as the equivalent clustering model is not robust to outliers. To overcome this limitation, we introduce a trimming function to the clustering model and propose a trimmed-clustering based l1-PCA (TC-PCA). With this trimming set formulation, the TC-PCA is not sensitive to outliers. Besides, we mathematically prove the convergence of the proposed algorithm. Experimental results on image classification and clustering indicate that our proposed method outperforms the current state-of-the-art methods.

[1]  Pascal Vincent,et al.  Representation Learning: A Review and New Perspectives , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[2]  P. Rousseeuw,et al.  A fast algorithm for the minimum covariance determinant estimator , 1999 .

[3]  Guoying Li,et al.  Breakdown properties of location $M$-estimators , 1998 .

[4]  Panos P. Markopoulos,et al.  Optimal Algorithms for L1-subspace Signal Processing , 2014, IEEE Transactions on Signal Processing.

[5]  Gentiane Haesbroeck,et al.  Outliers detection with the minimum covariance determinant estimator in practice , 2009 .

[6]  Alejandro F. Frangi,et al.  Two-dimensional PCA: a new approach to appearance-based face representation and recognition , 2004 .

[7]  Constantine Caramanis,et al.  Robust PCA via Outlier Pursuit , 2010, IEEE Transactions on Information Theory.

[8]  Dit-Yan Yeung,et al.  Bayesian Robust Matrix Factorization for Image and Video Processing , 2013, 2013 IEEE International Conference on Computer Vision.

[9]  Xiaogang Wang,et al.  Two-Dimensional Maximum Local Variation Based on Image Euclidean Distance for Face Recognition , 2013, IEEE Transactions on Image Processing.

[10]  Feiping Nie,et al.  $\ell _{2,p}$ -Norm Based PCA for Image Recognition , 2018, IEEE Transactions on Image Processing.

[11]  Jingdong Wang,et al.  A Probabilistic Approach to Robust Matrix Factorization , 2012, ECCV.

[12]  Feiping Nie,et al.  Optimal Mean Robust Principal Component Analysis , 2014, ICML.

[13]  Karl Pearson F.R.S. LIII. On lines and planes of closest fit to systems of points in space , 1901 .

[14]  Chris H. Q. Ding,et al.  R1-PCA: rotational invariant L1-norm principal component analysis for robust subspace factorization , 2006, ICML.

[15]  Xiaogang Wang,et al.  A unified framework for subspace face recognition , 2004, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[16]  Li Zhang,et al.  Joint Low-Rank and Sparse Principal Feature Coding for Enhanced Robust Representation and Visual Classification , 2016, IEEE Transactions on Image Processing.

[17]  Javier Ruiz-del-Solar,et al.  Analysis and Comparison of Eigenspace-Based Face Recognition Approaches , 2002, Int. J. Pattern Recognit. Artif. Intell..

[18]  Zahir M. Hussain,et al.  A feature-based structural measure: an image similarity measure for face recognition , 2017 .

[19]  Lawrence Carin,et al.  Bayesian Robust Principal Component Analysis , 2011, IEEE Transactions on Image Processing.

[20]  J. Brooks,et al.  A Pure L1-norm Principal Component Analysis. , 2013, Computational statistics & data analysis.

[21]  John Wright,et al.  Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Matrices via Convex Optimization , 2009, NIPS.

[22]  Yongqiang Zhao,et al.  Total Variation and Rank-1 Constraint RPCA for Background Subtraction , 2018, IEEE Access.

[23]  Qianqian Wang,et al.  Two-Dimensional PCA with F-Norm Minimization , 2017, AAAI.

[24]  Jiashu Zhang,et al.  Linear Discriminant Analysis Based on L1-Norm Maximization , 2013, IEEE Transactions on Image Processing.

[25]  Ajmal S. Mian,et al.  Joint Group Sparse PCA for Compressed Hyperspectral Imaging , 2015, IEEE Transactions on Image Processing.

[26]  Chris H. Q. Ding,et al.  K-means clustering via principal component analysis , 2004, ICML.

[27]  Xinbo Gao,et al.  Dimensionality Reduction by Integrating Sparse Representation and Fisher Criterion and its Applications , 2015, IEEE Transactions on Image Processing.

[28]  Qinghua Zheng,et al.  Avoiding Optimal Mean Robust PCA/2DPCA with Non-greedy ℓ1-Norm Maximization , 2016, IJCAI.

[29]  Ig-Jae Kim,et al.  PSI-CNN: A Pyramid-Based Scale-Invariant CNN Architecture for Face Recognition Robust to Various Image Resolutions , 2018, Applied Sciences.

[30]  Muhittin Gökmen,et al.  An Efficient Multiscale Scheme Using Local Zernike Moments for Face Recognition , 2018 .

[31]  Rong Wang,et al.  Robust 2DPCA With Non-greedy $\ell _{1}$ -Norm Maximization for Image Analysis , 2015, IEEE Transactions on Cybernetics.

[32]  Nojun Kwak,et al.  Principal Component Analysis by $L_{p}$ -Norm Maximization , 2014, IEEE Transactions on Cybernetics.

[33]  Lei Zhang,et al.  Robust Principal Component Analysis with Complex Noise , 2014, ICML.

[34]  Julie Josse,et al.  Selecting the number of components in principal component analysis using cross-validation approximations , 2012, Comput. Stat. Data Anal..

[35]  Panos P. Markopoulos,et al.  Efficient L1-Norm Principal-Component Analysis via Bit Flipping , 2016, IEEE Transactions on Signal Processing.

[36]  Nojun Kwak,et al.  Principal Component Analysis Based on L1-Norm Maximization , 2008, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[37]  Xuelong Li,et al.  L1-Norm-Based 2DPCA , 2010, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[38]  Dacheng Tao,et al.  Double Shrinking Sparse Dimension Reduction , 2013, IEEE Transactions on Image Processing.

[39]  Takeo Kanade,et al.  Robust L/sub 1/ norm factorization in the presence of outliers and missing data by alternative convex programming , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[40]  Jiajia Zhang,et al.  Principal components adjusted variable screening , 2017, Comput. Stat. Data Anal..

[41]  Deyu Meng,et al.  Divide-and-Conquer Method for L1 Norm Matrix Factorization in the Presence of Outliers and Missing Data , 2012, ArXiv.

[42]  Qinghua Zheng,et al.  Avoiding Optimal Mean ℓ2,1-Norm Maximization-Based Robust PCA for Reconstruction , 2017, Neural Computation.

[43]  Volkan Cevher,et al.  Bilinear Generalized Approximate Message Passing—Part II: Applications , 2014, IEEE Transactions on Signal Processing.

[44]  Ran He,et al.  Robust Principal Component Analysis Based on Maximum Correntropy Criterion , 2011, IEEE Transactions on Image Processing.

[45]  J. Tropp,et al.  Two proposals for robust PCA using semidefinite programming , 2010, 1012.1086.

[46]  Guangliang Chen,et al.  Simple, fast and accurate hyper-parameter tuning in Gaussian-kernel SVM , 2017, 2017 International Joint Conference on Neural Networks (IJCNN).

[47]  Yi Ma,et al.  Robust principal component analysis? , 2009, JACM.

[48]  Jian Yang,et al.  Nuclear Norm-Based 2-DPCA for Extracting Features From Images , 2015, IEEE Transactions on Neural Networks and Learning Systems.

[49]  Junbin Gao,et al.  Image Outlier Detection and Feature Extraction via L1-Norm-Based 2D Probabilistic PCA , 2015, IEEE Transactions on Image Processing.

[50]  Volkan Cevher,et al.  Bilinear Generalized Approximate Message Passing—Part I: Derivation , 2013, IEEE Transactions on Signal Processing.

[51]  Feiping Nie,et al.  A Non-Greedy Algorithm for L1-Norm LDA , 2017, IEEE Transactions on Image Processing.

[52]  Peter Meer,et al.  Generalized projection based M-estimator: Theory and applications , 2011, CVPR 2011.

[53]  Jian Yang,et al.  Sparse tensor discriminant analysis , 2013, IEEE Transactions on Image Processing.

[54]  Feiping Nie,et al.  Robust Principal Component Analysis with Non-Greedy l1-Norm Maximization , 2011, IJCAI.

[55]  Peter Meer,et al.  Generalized Projection-Based M-Estimator , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.