Interpretable Robust Feature Selection via Joint -Norms Minimization

Dimension reduction is a hot topic in data processing field. The challenge lies in how to find a suitable feature subset in low-dimensional space to accurately summarize the important information in high-dimensional space, rather than redundant information or noise. This requires the proposed model to reasonably explain the importance of features and be robust to noise. In order to solve this problem, this paper proposes an interpretable robust feature selection method, in which both the reconstruction error term and the regularization term are constrained by -norm. The reconstruction error term can capture samples corroded by noise, while the regular term automatically finds a group of discriminative features on relatively clean samples. Experimental results show the effectiveness of the proposed method, especially on noise data sets.

[1]  Zhenyu He,et al.  Adaptive Weighted Sparse Principal Component Analysis for Robust Unsupervised Feature Selection , 2020, IEEE Transactions on Neural Networks and Learning Systems.

[2]  Deng Cai,et al.  Laplacian Score for Feature Selection , 2005, NIPS.

[3]  Huan Liu,et al.  Spectral feature selection for supervised and unsupervised learning , 2007, ICML '07.

[4]  Xuelong Li,et al.  Unsupervised Feature Selection with Structured Graph Optimization , 2016, AAAI.

[5]  Navid Shokouhi,et al.  Sparse Principal Component Analysis With Preserved Sparsity Pattern , 2019, IEEE Transactions on Image Processing.

[6]  Chengqi Zhang,et al.  Convex Sparse PCA for Unsupervised Feature Learning , 2014, ACM Trans. Knowl. Discov. Data.

[7]  Massimiliano Pontil,et al.  Convex multi-task feature learning , 2008, Machine Learning.

[8]  Feiping Nie,et al.  Effective Discriminative Feature Selection With Nontrivial Solution , 2015, IEEE Transactions on Neural Networks and Learning Systems.

[9]  Carla E. Brodley,et al.  Feature Selection for Unsupervised Learning , 2004, J. Mach. Learn. Res..

[10]  Chris H. Q. Ding,et al.  R1-PCA: rotational invariant L1-norm principal component analysis for robust subspace factorization , 2006, ICML.

[11]  Songcan Chen,et al.  A unified algorithm for mixed $$l_{2,p}$$l2,p-minimizations and its application in feature selection , 2014, Comput. Optim. Appl..

[12]  Zi Huang,et al.  Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence ℓ2,1-Norm Regularized Discriminative Feature Selection for Unsupervised Learning , 2022 .

[13]  Lei Wang,et al.  Efficient Spectral Feature Selection with Minimum Redundancy , 2010, AAAI.

[14]  Feiping Nie,et al.  Trace Ratio Criterion for Feature Selection , 2008, AAAI.