Two-Dimensional Semi-Supervised Feature Selection

Data with a few labels are common in reality and corresponding semi-supervised learning methods have good application percepts such as image process and computer vision. However, most of conventional semi-supervised feature selection methods, they first convert the image data into corresponding vectors and then utilize the vectors for feature selection. However, these vector-based methods fail to consider the intrinsic location information of the matrix data resulting in degrading performance. To overcome these problems, in this paper, we propose a two-dimensional semi-supervised feature selection model, which directly selects the features from the matrix data by using both labeled and unlabeled data. Specifically, our proposed method unifies the sparse matrix regression and label prediction procedures to conduct feature selection. To solve our model, an alternating optimization algorithm is also proposed. Furthermore, we provide analysis about our proposed model including convergence analysis and computational complexity. Extensive experimental results confirm the effectiveness of the proposed model.

[1]  Feiping Nie,et al.  Semi-supervised Feature Selection via Rescaled Linear Regression , 2017, IJCAI.

[2]  Sameer A. Nene,et al.  Columbia Object Image Library (COIL100) , 1996 .

[3]  Jian Yang,et al.  A General Exponential Framework for Dimensionality Reduction , 2014, IEEE Transactions on Image Processing.

[4]  Yiu-ming Cheung,et al.  Feature Selection and Kernel Learning for Local Learning-Based Clustering , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[5]  Xiaofei He,et al.  Locality Preserving Projections , 2003, NIPS.

[6]  Feiping Nie,et al.  2D Feature Selection by Sparse Matrix Regression , 2017, IEEE Transactions on Image Processing.

[7]  Jiawei Han,et al.  Generalized Fisher Score for Feature Selection , 2011, UAI.

[8]  Feiping Nie,et al.  Discriminative Least Squares Regression for Multiclass Classification and Feature Selection , 2012, IEEE Transactions on Neural Networks and Learning Systems.

[9]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[10]  Avinash C. Kak,et al.  PCA versus LDA , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[11]  Zenglin Xu,et al.  Discriminative Semi-Supervised Feature Selection Via Manifold Regularization , 2009, IEEE Transactions on Neural Networks.

[12]  Yuan Yan Tang,et al.  Joint sparse matrix regression and nonnegative spectral analysis for two-dimensional unsupervised feature selection , 2019, Pattern Recognit..

[13]  W. Krzanowski Selection of Variables to Preserve Multivariate Data Structure, Using Principal Components , 1987 .

[14]  Stephen Lin,et al.  Graph Embedding and Extensions: A General Framework for Dimensionality Reduction , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[15]  Deng Cai,et al.  Laplacian Score for Feature Selection , 2005, NIPS.

[16]  Yuan Yan Tang,et al.  Graph-based multiple rank regression for image classification , 2018, Neurocomputing.

[17]  Feiping Nie,et al.  A Self-Balanced Min-Cut Algorithm for Image Clustering , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[18]  Jing Liu,et al.  Unsupervised Feature Selection Using Nonnegative Spectral Analysis , 2012, AAAI.

[19]  Simon C. K. Shiu,et al.  Unsupervised feature selection by regularized self-representation , 2015, Pattern Recognit..

[20]  Jian Yang,et al.  Two-dimensional PCA: a new approach to appearance-based face representation and recognition , 2004, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[21]  Jieping Ye,et al.  Two-Dimensional Linear Discriminant Analysis , 2004, NIPS.

[22]  Yuxiao Hu,et al.  Face recognition using Laplacianfaces , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[23]  Zoubin Ghahramani,et al.  Combining active learning and semi-supervised learning using Gaussian fields and harmonic functions , 2003, ICML 2003.

[24]  Fuhui Long,et al.  Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy , 2003, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[25]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[26]  Michael J. Lyons,et al.  Automatic Classification of Single Facial Images , 1999, IEEE Trans. Pattern Anal. Mach. Intell..

[27]  Xuelong Li,et al.  Joint Embedding Learning and Sparse Regression: A Framework for Unsupervised Feature Selection , 2014, IEEE Transactions on Cybernetics.

[28]  Feiping Nie,et al.  Efficient and Robust Feature Selection via Joint ℓ2, 1-Norms Minimization , 2010, NIPS.

[29]  Bin Luo,et al.  2D-LPP: A two-dimensional extension of locality preserving projections , 2007, Neurocomputing.