Deep random walk of unitary invariance for large-scale data representation

Abstract Data representation aims at learning an efficient low-dimensional representation, which is always a challenging task in machine learning and computer vision. It can largely improve the performance of specific learning tasks. Unsupervised methods are extensively applied to data representation, which considers the internal connection among data. Most of existing unsupervised models usually use a specific norm to favor certain distributions of the input data, leading to an unsustainable encouraging performance for given learning tasks. In this paper, we propose an efficient data representation method to address large-scale feature representation problems, where the deep random walk of unitary invariance is exploited for learning discriminative features. First, the data representation is formulated as deep random walk problems, where unitarily invariant norms are employed to capture diverse beneficial perspectives hidden in the data. It is embedded into a state transition matrix model, where an arbitrary number of transition steps is available for an accurate affinity evaluation. Second, data representation problems are then transformed as high-order matrix factorization tasks with unitary invariance. Third, a closed-form solution is proved for the formulated data representation problem, which may provide a new perspective for solving high-order matrix factorization problems. Finally, extensive comparative experiments are conducted in publicly available real-world data sets. In addition, experimental results demonstrate that the proposed method achieves better performance than other compared state-of-the-art approaches in terms of data clustering.

[1]  Jing Liu,et al.  Clustering-Guided Sparse Structural Learning for Unsupervised Feature Selection , 2014, IEEE Transactions on Knowledge and Data Engineering.

[2]  Feiping Nie,et al.  Optimal Mean Robust Principal Component Analysis , 2014, ICML.

[3]  Witold Pedrycz,et al.  Subspace learning for unsupervised feature selection via matrix factorization , 2015, Pattern Recognit..

[4]  Liqiang Nie,et al.  Learning robust affinity graph representation for multi-view clustering , 2021, Inf. Sci..

[5]  Chia-Wen Lin,et al.  CNN-Based Joint Clustering and Representation Learning with Feature Drift Compensation for Large-Scale Image Data , 2017, IEEE Transactions on Multimedia.

[6]  Karl Stratos,et al.  Discrete Latent Variable Representations for Low-Resource Text Classification , 2020, ACL.

[7]  H. Sebastian Seung,et al.  Learning the parts of objects by non-negative matrix factorization , 1999, Nature.

[8]  Biao Huang,et al.  Deep Learning-Based Feature Representation and Its Application for Soft Sensor Modeling With Variable-Wise Weighted SAE , 2018, IEEE Transactions on Industrial Informatics.

[9]  Sen Wu,et al.  Robust sparse coding via self-paced learning for data representation , 2021, Inf. Sci..

[10]  Shuicheng Yan,et al.  Neighborhood preserving embedding , 2005, Tenth IEEE International Conference on Computer Vision (ICCV'05) Volume 1.

[11]  Fuhui Long,et al.  Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy , 2003, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[12]  Jing Liu,et al.  Robust Structured Subspace Learning for Data Representation , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[13]  Joel Z. Leibo,et al.  Prefrontal cortex as a meta-reinforcement learning system , 2018, bioRxiv.

[14]  Qi Zhu,et al.  Adaptive feature weighting for robust Lp-norm sparse representation with application to biometric image classification , 2020, Int. J. Mach. Learn. Cybern..

[15]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[16]  Zeynep Akata,et al.  Learning Robust Representations via Multi-View Information Bottleneck , 2020, ICLR.

[17]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[18]  Deli Zhao,et al.  Network Representation Learning with Rich Text Information , 2015, IJCAI.

[19]  Yun Fu,et al.  Discerning Feature Supported Encoder for Image Representation , 2019, IEEE Transactions on Image Processing.

[20]  Xuelong Li,et al.  Canonical Correlation Analysis With L2,1-Norm for Multiview Data Representation , 2020, IEEE Transactions on Cybernetics.

[21]  Wenzhong Guo,et al.  Sparse Multigraph Embedding for Multimodal Feature Representation , 2017, IEEE Transactions on Multimedia.

[22]  Zhiyuan Liu,et al.  Max-Margin DeepWalk: Discriminative Learning of Network Representation , 2016, IJCAI.

[23]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[24]  Wei Liu,et al.  Tensor Robust Principal Component Analysis with a New Tensor Nuclear Norm , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[25]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[26]  Mikhail Belkin,et al.  Laplacian Eigenmaps and Spectral Techniques for Embedding and Clustering , 2001, NIPS.

[27]  Chen Gong,et al.  Multi-Manifold Positive and Unlabeled Learning for Visual Analysis , 2020, IEEE Transactions on Circuits and Systems for Video Technology.

[28]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[29]  Yong-Sheng Chen,et al.  MVSNet++: Learning Depth-Based Attention Pyramid Features for Multi-View Stereo , 2020, IEEE Transactions on Image Processing.

[30]  Peng Li,et al.  Rank-constrained nonnegative matrix factorization for data representation , 2020, Inf. Sci..

[31]  Stefano Ermon,et al.  Tile2Vec: Unsupervised representation learning for spatially distributed data , 2018, AAAI.

[32]  John Shawe-Taylor,et al.  Canonical Correlation Analysis: An Overview with Application to Learning Methods , 2004, Neural Computation.

[33]  Wenyin Liu,et al.  A novel feature representation: Aggregating convolution kernels for image retrieval , 2020, Neural Networks.

[34]  Murat Ekinci,et al.  Single image super resolution using dictionary learning and sparse coding with multi-scale and multi-directional Gabor feature representation , 2020, Inf. Sci..

[35]  Shuyuan Yang,et al.  Semi-Supervised Graph Regularized Deep NMF With Bi-Orthogonal Constraints for Data Representation , 2020, IEEE Transactions on Neural Networks and Learning Systems.

[36]  Dong-Seong Kim,et al.  Image representation of pose-transition feature for 3D skeleton-based action recognition , 2020, Inf. Sci..

[37]  Wenyin Liu,et al.  Shared Multi-View Data Representation for Multi-Domain Event Detection , 2020, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[38]  Huaijiang Sun,et al.  Joint dimensionality reduction and metric learning for image set classification , 2020, Inf. Sci..

[39]  Vishal M. Patel,et al.  Generative-Discriminative Feature Representations for Open-Set Recognition , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[40]  Xuelong Li,et al.  Horizontal and Vertical Nuclear Norm-Based 2DLDA for Image Representation , 2019, IEEE Transactions on Circuits and Systems for Video Technology.

[41]  Fei Li,et al.  Citation Recommendation with a Content-Sensitive DeepWalk Based Approach , 2019, 2019 International Conference on Data Mining Workshops (ICDMW).

[42]  Yaoliang Yu,et al.  Rank/Norm Regularization with Closed-Form Solutions: Application to Subspace Clustering , 2011, UAI.

[43]  Xiaojie Su,et al.  Robust Manhattan non-negative matrix factorization for image recovery and representation , 2020, Inf. Sci..

[44]  William Zhu,et al.  Sparse Graph Embedding Unsupervised Feature Selection , 2018, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[45]  Tao Mei,et al.  Deep Collaborative Embedding for Social Image Understanding , 2019, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[46]  Ilya Feige Invariant-equivariant representation learning for multi-class data , 2019, ICML.

[47]  Jian Yang,et al.  KPCA plus LDA: a complete kernel Fisher discriminant framework for feature extraction and recognition , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[48]  Rui Li,et al.  Link prediction based on feature representation and fusion , 2021, Inf. Sci..

[49]  Chao Chen,et al.  Joint Domain Alignment and Discriminative Feature Learning for Unsupervised Deep Domain Adaptation , 2018, AAAI.

[50]  Feiping Nie,et al.  Discriminative Unsupervised Dimensionality Reduction , 2015, IJCAI.

[51]  Ming Gu,et al.  Randomized Projection for Rank-Revealing Matrix Factorizations and Low-Rank Approximations , 2020, SIAM Rev..