Similarity-Adaptive Latent Low-Rank Representation for Robust Data Representation

We propose a novel Similarity-Adaptive Latent Low-Rank Representation (SA-LatLRR) model for the robust representation and subspace recovery. SA-LatLRR inherits all merits of recent LatLRR, and further improves it by enhancing the representations. SA-LatLRR aims at decomposing given data into a principal feature part encoded by the Frobenius-norm based coefficients, a similarity-adaptive salient feature part and a sparse error part. Specifically, our SA-LatLRR incorporates a reconstructive error minimization term over coefficients and salient features, which can clearly preserve the neighborhood information of salient features adaptively. The added regularization can also encourage the coefficients to be block-diagonal and discriminative, as the shared coefficients could minimize the reconstruction errors over both original data and salient features at the same time, where the embedded salient features contain less noise and unfavorable features than the original data. Moreover, to make salient features more informative and robust to noise, SA-LatLRR imposes the sparse L2,1-norm and low-rank constraints on the projection jointly so that the features are more notable and discriminative. The Frobenius-norm based principal feature part can also make the coefficients coding process very efficient. Extensive comparison results demonstrate the validity of our SA-LatLRR.

[1]  Krishnakumar Balasubramanian,et al.  Smooth sparse coding via marginal regression for learning sparse representations , 2012, Artif. Intell..

[2]  Qingshan Liu,et al.  A Deterministic Analysis for LRR , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[3]  Shiyu Chang,et al.  Low-Rank Sparse Feature Selection for Patient Similarity Learning , 2016, 2016 IEEE 16th International Conference on Data Mining (ICDM).

[4]  Shuicheng Yan,et al.  Latent Low-Rank Representation for subspace segmentation and feature extraction , 2011, 2011 International Conference on Computer Vision.

[5]  Yan Zhang,et al.  Discriminative sparse flexible manifold embedding with novel graph for robust visual representation and label propagation , 2017, Pattern Recognit..

[6]  Jiang Li,et al.  Dimensionality reduction of hyperspectral data using discrete wavelet transform feature extraction , 2002, IEEE Trans. Geosci. Remote. Sens..

[7]  David D. Cox,et al.  Making a Science of Model Search: Hyperparameter Optimization in Hundreds of Dimensions for Vision Architectures , 2013, ICML.

[8]  Bingbing Ni,et al.  Multitask Low-Rank Affinity Graph for Image Segmentation and Image Annotation , 2016, ACM Trans. Intell. Syst. Technol..

[9]  Xu-Dong Zhang,et al.  Learning to Rank from Noisy Data , 2015, ACM Trans. Intell. Syst. Technol..

[10]  Shuicheng Yan,et al.  Bilinear low-rank coding framework and extension for robust image recovery and feature representation , 2015, Knowl. Based Syst..

[11]  Yulong Wang,et al.  Graph-Regularized Low-Rank Representation for Destriping of Hyperspectral Images , 2013, IEEE Transactions on Geoscience and Remote Sensing.

[12]  Xuelong Li,et al.  Joint Embedding Learning and Sparse Regression: A Framework for Unsupervised Feature Selection , 2014, IEEE Transactions on Cybernetics.

[13]  Tommy W. S. Chow,et al.  Binary- and Multi-class Group Sparse Canonical Correlation Analysis for Feature Extraction and Classification , 2013, IEEE Transactions on Knowledge and Data Engineering.

[14]  Feiping Nie,et al.  Efficient and Robust Feature Selection via Joint ℓ2, 1-Norms Minimization , 2010, NIPS.

[15]  Guillermo Sapiro,et al.  Discriminative learned dictionaries for local image analysis , 2008, 2008 IEEE Conference on Computer Vision and Pattern Recognition.

[16]  Li Zhang,et al.  Joint Low-Rank and Sparse Principal Feature Coding for Enhanced Robust Representation and Visual Classification , 2016, IEEE Transactions on Image Processing.

[17]  D. B. Graham,et al.  Characterising Virtual Eigensignatures for General Purpose Face Recognition , 1998 .

[18]  Sijia Cai,et al.  Accelerated matrix recovery via random projection based on inexact augmented Lagrange multiplier method , 2013 .

[19]  G. Sapiro,et al.  A collaborative framework for 3D alignment and classification of heterogeneous subvolumes in cryo-electron tomography. , 2013, Journal of structural biology.

[20]  Xiaofei He,et al.  Locality Preserving Projections , 2003, NIPS.

[21]  Ting Wang,et al.  Kernel Sparse Representation-Based Classifier , 2012, IEEE Transactions on Signal Processing.

[22]  Yong Yu,et al.  Robust Recovery of Subspace Structures by Low-Rank Representation , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[23]  Mia Hubert,et al.  ROBPCA: A New Approach to Robust Principal Component Analysis , 2005, Technometrics.

[24]  T. Kanade,et al.  Combining Models and Exemplars for Face Recognition: An Illuminating Example , 2001 .

[25]  Jiawei Han,et al.  Isometric Projection , 2007, AAAI.

[26]  Guillermo Sapiro,et al.  Sparse Representation for Computer Vision and Pattern Recognition , 2010, Proceedings of the IEEE.

[27]  Kuldip K. Paliwal,et al.  Feature extraction and dimensionality reduction algorithms and their applications in vowel recognition , 2003, Pattern Recognit..

[28]  Shuicheng Yan,et al.  Similarity preserving low-rank representation for enhanced data representation and effective subspace learning , 2014, Neural Networks.

[29]  Changsheng Xu,et al.  Inductive Robust Principal Component Analysis , 2012, IEEE Transactions on Image Processing.

[30]  John Wright,et al.  Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Matrices via Convex Optimization , 2009, NIPS.

[31]  Yong Yu,et al.  Robust Subspace Segmentation by Low-Rank Representation , 2010, ICML.

[32]  Yi Ma,et al.  Robust principal component analysis? , 2009, JACM.

[33]  Jürgen Schmidhuber,et al.  Deep learning in neural networks: An overview , 2014, Neural Networks.