Feature Extraction for Incomplete Data Via Low-Rank Tensor Decomposition With Feature Regularization

Multidimensional data (i.e., tensors) with missing entries are common in practice. Extracting features from incomplete tensors is an important yet challenging problem in many fields such as machine learning, pattern recognition, and computer vision. Although the missing entries can be recovered by tensor completion techniques, these completion methods focus only on missing data estimation instead of effective feature extraction. To the best of our knowledge, the problem of feature extraction from incomplete tensors has yet to be well explored in the literature. In this paper, we therefore tackle this problem within the unsupervised learning environment. Specifically, we incorporate low-rank tensor decomposition with feature variance maximization (TDVM) in a unified framework. Based on orthogonal Tucker and CP decompositions, we design two TDVM methods, TDVM-Tucker and TDVM-CP, to learn low-dimensional features viewing the core tensors of the Tucker model as features and viewing the weight vectors of the CP model as features. TDVM explores the relationship among data samples via maximizing feature variance and simultaneously estimates the missing entries via low-rank Tucker/CP approximation, leading to informative features extracted directly from observed entries. Furthermore, we generalize the proposed methods by formulating a general model that incorporates feature regularization into low-rank tensor approximation. In addition, we develop a joint optimization scheme to solve the proposed methods by integrating the alternating direction method of multipliers with the block coordinate descent method. Finally, we evaluate our methods on six real-world image and video data sets under a newly designed multiblock missing setting. The extracted features are evaluated in face recognition, object/action classification, and face/gait clustering. Experimental results demonstrate the superior performance of the proposed methods compared with the state-of-the-art approaches.

[1]  Robert P. Goldman,et al.  Imputation of Missing Data Using Machine Learning Techniques , 1996, KDD.

[2]  Jian Yang,et al.  Modified Principal Component Analysis: An Integration of Multiple Similarity Subspace Models , 2014, IEEE Transactions on Neural Networks and Learning Systems.

[3]  Wei Chu,et al.  Probabilistic Models for Incomplete Multi-dimensional Arrays , 2009, AISTATS.

[4]  Guoqiang Zhong,et al.  Low-Rank Tensor Learning with Discriminant Analysis for Action Classification and Image Recovery , 2014, AAAI.

[5]  Sameer A. Nene,et al.  Columbia Object Image Library (COIL100) , 1996 .

[6]  Wei Liu,et al.  Tensor Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Tensors via Convex Optimization , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  J. Chang,et al.  Analysis of individual differences in multidimensional scaling via an n-way generalization of “Eckart-Young” decomposition , 1970 .

[8]  Zemin Zhang,et al.  Exact Tensor Completion Using t-SVD , 2015, IEEE Transactions on Signal Processing.

[9]  Jieping Ye,et al.  Tensor Completion for Estimating Missing Values in Visual Data , 2013, IEEE Trans. Pattern Anal. Mach. Intell..

[10]  Jing Zhang,et al.  Low-Rank Regularized Heterogeneous Tensor Decomposition for Subspace Clustering , 2018, IEEE Signal Processing Letters.

[11]  Bin Dong,et al.  Fast Linearized Bregman Iteration for Compressive Sensing and Sparse Denoising , 2011, ArXiv.

[12]  Zhang Yi,et al.  Constructing the L2-Graph for Robust Subspace Learning and Subspace Clustering , 2012, IEEE Transactions on Cybernetics.

[13]  Prateek Jain,et al.  Provable Tensor Factorization with Missing Data , 2014, NIPS.

[14]  Joos Vandewalle,et al.  A Multilinear Singular Value Decomposition , 2000, SIAM J. Matrix Anal. Appl..

[15]  Vin de Silva,et al.  Tensor rank and the ill-posedness of the best low-rank approximation problem , 2006, math/0607647.

[16]  Qiquan Shi,et al.  Tensor Rank Estimation and Completion via CP-based Nuclear Norm , 2017, CIKM.

[17]  Liqing Zhang,et al.  Bayesian CP Factorization of Incomplete Tensors with Automatic Rank Determination , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[18]  Anima Anandkumar,et al.  Tensor decompositions for learning latent variable models , 2012, J. Mach. Learn. Res..

[19]  J. Denis,et al.  Orthogonal tensor decomposition of 3-way tables , 1989 .

[20]  Jicong Fan,et al.  Sparse subspace clustering for data with missing entries and high-rank matrix completion , 2017, Neural Networks.

[21]  Junbin Gao,et al.  Vectorial Dimension Reduction for Tensors Based on Bayesian Inference , 2017, IEEE Transactions on Neural Networks and Learning Systems.

[22]  Ronen Basri,et al.  Actions as Space-Time Shapes , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[23]  Kenneth Lange,et al.  Matrix completion discriminant analysis , 2015, Comput. Stat. Data Anal..

[24]  Michel Verleysen,et al.  Feature selection with missing data using mutual information estimators , 2012, Neurocomputing.

[25]  Yandong Tang,et al.  A Generalized Model for Robust Tensor Factorization With Noise Modeling by Mixture of Gaussians , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[26]  Robert D. Nowak,et al.  High-Rank Matrix Completion , 2012, AISTATS.

[27]  Hyeonjoon Moon,et al.  The FERET evaluation methodology for face-recognition algorithms , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[28]  Hong Cheng,et al.  Generalized Higher Order Orthogonal Iteration for Tensor Learning and Decomposition , 2016, IEEE Transactions on Neural Networks and Learning Systems.

[29]  Misha Elena Kilmer,et al.  Novel Methods for Multilinear Data Completion and De-noising Based on Tensor-SVD , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[30]  Ehsan Elhamifar,et al.  High-Rank Matrix Completion and Clustering under Self-Expressive Models , 2016, NIPS.

[31]  Andrzej Cichocki,et al.  Efficient Nonnegative Tucker Decompositions: Algorithms and Uniqueness , 2014, IEEE Transactions on Image Processing.

[32]  Ivan Oseledets,et al.  Tensor-Train Decomposition , 2011, SIAM J. Sci. Comput..

[33]  R. Vidal,et al.  Sparse Subspace Clustering: Algorithm, Theory, and Applications. , 2013, IEEE transactions on pattern analysis and machine intelligence.

[34]  Sudeep Sarkar,et al.  The humanID gait challenge problem: data sets, performance, and analysis , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[35]  Yuanyuan Liu,et al.  Generalized Higher-Order Tensor Decomposition via Parallel ADMM , 2014, AAAI.

[36]  Lawrence Carin,et al.  Incomplete-data classification using logistic regression , 2005, ICML.

[37]  Lawrence Carin,et al.  On Classification with Incomplete Data , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[38]  Yunming Ye,et al.  MR-NTD: Manifold Regularization Nonnegative Tucker Decomposition for Tensor Data Dimension Reduction and Representation , 2017, IEEE Transactions on Neural Networks and Learning Systems.

[39]  Huseyin Ozkan,et al.  Data Imputation Through the Identification of Local Anomalies , 2015, IEEE Transactions on Neural Networks and Learning Systems.

[40]  Shuicheng Yan,et al.  Latent Low-Rank Representation for subspace segmentation and feature extraction , 2011, 2011 International Conference on Computer Vision.

[41]  Haiping Lu,et al.  Bilinear Probabilistic Canonical Correlation Analysis via Hybrid Concatenations , 2017, AAAI.

[42]  Roi Livni,et al.  Classification with Low Rank and Missing Data , 2015, ICML.

[43]  Gene H. Golub,et al.  Rank-One Approximation to High Order Tensors , 2001, SIAM J. Matrix Anal. Appl..

[44]  Erik G. Larsson,et al.  The Higher-Order Singular Value Decomposition: Theory and an Application [Lecture Notes] , 2010, IEEE Signal Processing Magazine.

[45]  Richard A. Harshman,et al.  Foundations of the PARAFAC procedure: Models and conditions for an "explanatory" multi-model factor analysis , 1970 .

[46]  Qiquan Shi,et al.  Semi-Orthogonal Multilinear PCA with Relaxed Start , 2015, IJCAI.

[47]  Haiping Lu,et al.  Uncorrelated Multilinear Principal Component Analysis for Unsupervised Multilinear Subspace Learning , 2009, IEEE Transactions on Neural Networks.

[48]  Tamara G. Kolda,et al.  Tensor Decompositions and Applications , 2009, SIAM Rev..

[49]  Ian T. Jolliffe,et al.  Principal Component Analysis , 2002, International Encyclopedia of Statistical Science.

[50]  Wensheng Zhang,et al.  The Twist Tensor Nuclear Norm for Video Completion , 2017, IEEE Transactions on Neural Networks and Learning Systems.

[51]  相原 龍,et al.  Alternating Direction Method of Multipliersを用いた声質変換のためのパラレル辞書学習 (音声) -- (第17回音声言語シンポジウム) , 2015 .

[52]  Emmanuel J. Candès,et al.  A Singular Value Thresholding Algorithm for Matrix Completion , 2008, SIAM J. Optim..

[53]  Qiquan Shi,et al.  Feature Extraction for Incomplete Data via Low-rank Tucker Decomposition , 2017, ECML/PKDD.

[54]  Minh N. Do,et al.  Efficient Tensor Completion for Color Image and Video Recovery: Low-Rank Tensor Train , 2016, IEEE Transactions on Image Processing.

[55]  Aníbal R. Figueiras-Vidal,et al.  Pattern classification with missing data: a review , 2010, Neural Computing and Applications.

[56]  Aleix M. Martinez,et al.  The AR face database , 1998 .

[57]  Daniel P. Robinson,et al.  Sparse Subspace Clustering with Missing Entries , 2015, ICML.

[58]  Philip S. Yu,et al.  t-BNE: Tensor-based Brain Network Embedding , 2017, SDM.

[59]  Yang Zhou,et al.  Probabilistic Rank-One Discriminant Analysis via Collective and Individual Variation Modeling , 2020, IEEE Transactions on Cybernetics.

[60]  Emmanuel J. Candès,et al.  Exact Matrix Completion via Convex Optimization , 2009, Found. Comput. Math..

[61]  Haiping Lu,et al.  MPCA: Multilinear Principal Component Analysis of Tensor Objects , 2008, IEEE Transactions on Neural Networks.

[62]  Hong Cheng,et al.  Trace Norm Regularized CANDECOMP/PARAFAC Decomposition With Missing Data , 2015, IEEE Transactions on Cybernetics.

[63]  David J. Kriegman,et al.  Acquiring linear subspaces for face recognition under variable lighting , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[64]  Tamara G. Kolda,et al.  Scalable Tensor Factorizations for Incomplete Data , 2010, ArXiv.

[65]  Jiwen Lu,et al.  Automatic Subspace Learning via Principal Coefficients Embedding , 2014, IEEE Transactions on Cybernetics.

[66]  Yiu-Ming Cheung,et al.  Rank-One Matrix Completion With Automatic Rank Estimation via L1-Norm Regularization , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[67]  Hong Cheng,et al.  Generalized Higher-Order Orthogonal Iteration for Tensor Decomposition and Completion , 2014, NIPS.

[68]  Allen Y. Yang,et al.  Robust Face Recognition via Sparse Representation , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.