Robust sparse coding via self-paced learning for data representation

Abstract Sparse coding (SC), due to its thorough theoretical property and outstanding effectiveness, is attracting more and more attention in various data representation and data mining applications. However, the optimization of most existing sparse coding algorithms are non-convex and thus prone to become stuck into bad local minima under the framework of alternative optimization, especially when there are many outliers and noisy data. To enhance the learning robustness, in this study, we will present an unified framework named Self-Paced Sparse Coding (SPSC), which gradually includes data into the learning process of SC from easy ones to complex ones by incorporating self-paced learning methodology. It implements a soft instance selection accordingly rather than a heuristic hard strategy sample selection. We also generalize the self-paced learning schema into different levels of dynamic selection on instances, features and elements respectively. Further, we show an optimization algorithm to solve it and a theoretical explanation to analyze the effectiveness of it. Extensive experimental results on the real-world clean image datasets and images with two kinds of corruptions demonstrate the remarkable robustness of the proposed method for high dimensional data representation on image clustering and reconstruction tasks over the state-of-the-arts.

[1]  Trac D. Tran,et al.  Structured Priors for Sparse-Representation-Based Hyperspectral Image Classification , 2014, IEEE Geoscience and Remote Sensing Letters.

[2]  Jane You,et al.  Image clustering by hyper-graph regularized non-negative matrix factorization , 2014, Neurocomputing.

[3]  Nicu Sebe,et al.  Multi-Paced Dictionary Learning for cross-domain retrieval and recognition , 2016, 2016 23rd International Conference on Pattern Recognition (ICPR).

[4]  Heng Huang,et al.  Self-Paced Network Embedding , 2018, KDD.

[5]  Dacheng Tao,et al.  Multi-view Self-Paced Learning for Clustering , 2015, IJCAI.

[6]  Shuiwang Ji,et al.  SLEP: Sparse Learning with Efficient Projections , 2011 .

[7]  Deyu Meng,et al.  A theoretical understanding of self-paced learning , 2017, Inf. Sci..

[8]  Jinhui Tang,et al.  Robust Structured Nonnegative Matrix Factorization for Image Representation , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[9]  Yan Zhang,et al.  Discrete Ranking-based Matrix Factorization with Self-Paced Learning , 2018, KDD.

[10]  Sumit Basu,et al.  Teaching Classification Boundaries to Humans , 2013, AAAI.

[11]  Jason Weston,et al.  Curriculum learning , 2009, ICML '09.

[12]  Qingshan Liu,et al.  Image retrieval via probabilistic hypergraph ranking , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[13]  Meng Wang,et al.  Disease Inference from Health-Related Questions via Sparse Deep Learning , 2015, IEEE Transactions on Knowledge and Data Engineering.

[14]  Liang-Tien Chia,et al.  Laplacian Sparse Coding, Hypergraph Laplacian Sparse Coding, and Applications , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[15]  Zhang Yi,et al.  Constructing the L2-Graph for Robust Subspace Learning and Subspace Clustering , 2012, IEEE Transactions on Cybernetics.

[16]  Wenjun Zhou,et al.  Multi-Hypergraph Consistent Sparse Coding , 2017, ACM Trans. Intell. Syst. Technol..

[17]  F. Vaida PARAMETER CONVERGENCE FOR EM AND MM ALGORITHMS , 2005 .

[18]  David Zhang,et al.  Fisher Discrimination Dictionary Learning for sparse representation , 2011, 2011 International Conference on Computer Vision.

[19]  Larry S. Davis,et al.  Label Consistent K-SVD: Learning a Discriminative Dictionary for Recognition , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[20]  Zhiyuan Liu,et al.  Neural Relation Extraction with Selective Attention over Instances , 2016, ACL.

[21]  Yang Gao,et al.  Self-paced dictionary learning for image classification , 2012, ACM Multimedia.

[22]  David J. Field,et al.  Emergence of simple-cell receptive field properties by learning a sparse code for natural images , 1996, Nature.

[23]  Dong Xu,et al.  Hyperplane-based nonnegative matrix factorization with label information , 2019, Inf. Sci..

[24]  Chun Chen,et al.  Graph Regularized Sparse Coding for Image Representation , 2011, IEEE Transactions on Image Processing.

[25]  Jiangping Wang,et al.  Data Clustering by Laplacian Regularized L1-Graph , 2014, AAAI.

[26]  R. Tibshirani,et al.  REJOINDER TO "LEAST ANGLE REGRESSION" BY EFRON ET AL. , 2004, math/0406474.

[27]  Fei Wang,et al.  Graph dual regularization non-negative matrix factorization for co-clustering , 2012, Pattern Recognit..

[28]  David A. Cohn,et al.  Active Learning with Statistical Models , 1996, NIPS.

[29]  Guangming Shi,et al.  Image restoration via Bayesian structured sparse coding , 2014, 2014 IEEE International Conference on Image Processing (ICIP).

[30]  Allen Y. Yang,et al.  Robust Face Recognition via Sparse Representation , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[31]  Shiguang Shan,et al.  Self-Paced Curriculum Learning , 2015, AAAI.

[32]  Allen Y. Yang,et al.  Fast ℓ1-minimization algorithms and an application in robust face recognition: A review , 2010, 2010 IEEE International Conference on Image Processing.

[33]  Jian Zhang,et al.  Convolutional Sparse Autoencoders for Image Classification , 2018, IEEE Transactions on Neural Networks and Learning Systems.