Matrix Co-completion for Multi-label Classification with Missing Features and Labels

We consider a challenging multi-label classification problem where both feature matrix $\X$ and label matrix $\Y$ have missing entries. An existing method concatenated $\X$ and $\Y$ as $[\X; \Y]$ and applied a matrix completion (MC) method to fill the missing entries, under the assumption that $[\X; \Y]$ is of low-rank. However, since entries of $\Y$ take binary values in the multi-label setting, it is unlikely that $\Y$ is of low-rank. Moreover, such assumption implies a linear relationship between $\X$ and $\Y$ which may not hold in practice. In this paper, we consider a latent matrix $\Z$ that produces the probability $\sigma(Z_{ij})$ of generating label $Y_{ij}$, where $\sigma(\cdot)$ is nonlinear. Considering label correlation, we assume $[\X; \Z]$ is of low-rank, and propose an MC algorithm based on subgradient descent named co-completion (COCO) motivated by elastic net and one-bit MC. We give a theoretical bound on the recovery effect of COCO and demonstrate its practical usefulness through experiments.

[1]  Weixiong Zhang,et al.  Marginalized Denoising for Link Prediction and Multi-Label Learning , 2015, AAAI.

[2]  Ambuj Tewari,et al.  Stochastic methods for l1 regularized loss minimization , 2009, ICML '09.

[3]  Mark Herbster,et al.  Mistake Bounds for Binary Matrix Completion , 2016, NIPS.

[4]  Emmanuel J. Candès,et al.  Matrix Completion With Noise , 2009, Proceedings of the IEEE.

[5]  G. Watson Characterization of the subdifferential of some matrix norms , 1992 .

[6]  Philip Resnik,et al.  Learning a Concept Hierarchy from Multi-labeled Documents , 2014, NIPS.

[7]  Rong Jin,et al.  Stochastic Proximal Gradient Descent for Nuclear Norm Regularization , 2015, ArXiv.

[8]  Zhi-Hua Zhou,et al.  Multilabel Neural Networks with Applications to Functional Genomics and Text Categorization , 2006, IEEE Transactions on Knowledge and Data Engineering.

[9]  Y. Nesterov Gradient methods for minimizing composite objective function , 2007 .

[10]  Zhi-Hua Zhou,et al.  Fast Multi-Instance Multi-Label Learning , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[11]  Emmanuel J. Candès,et al.  Exact Matrix Completion via Convex Optimization , 2008, Found. Comput. Math..

[12]  Kilian Q. Weinberger,et al.  Fast Image Tagging , 2013, ICML.

[13]  Robert D. Nowak,et al.  Transduction with Matrix Completion: Three Birds with One Stone , 2010, NIPS.

[14]  Rong Jin,et al.  Multi-label learning with incomplete class assignments , 2011, CVPR 2011.

[15]  Jason Weston,et al.  A kernel method for multi-labelled classification , 2001, NIPS.

[16]  Benjamin Recht,et al.  A Simpler Approach to Matrix Completion , 2009, J. Mach. Learn. Res..

[17]  Emmanuel J. Candès,et al.  A Singular Value Thresholding Algorithm for Matrix Completion , 2008, SIAM J. Optim..

[18]  Zhi-Hua Zhou,et al.  Multi-label Learning , 2017, Encyclopedia of Machine Learning and Data Mining.

[19]  Miao Xu,et al.  Speedup Matrix Completion with Side Information: Application to Multi-Label Learning , 2013, NIPS.

[20]  John Langford,et al.  Sparse Online Learning via Truncated Gradient , 2008, NIPS.

[21]  Gert R. G. Lanckriet,et al.  Semantic Annotation and Retrieval of Music and Sound Effects , 2008, IEEE Transactions on Audio, Speech, and Language Processing.

[22]  Inderjit S. Dhillon,et al.  Large-scale Multi-label Learning with Missing Labels , 2013, ICML.

[23]  Zhi-Hua Zhou,et al.  A Unified View of Multi-Label Performance Measures , 2016, ICML.

[24]  Vikas Sindhwani,et al.  Efficient and Practical Stochastic Subgradient Descent for Nuclear Norm Regularization , 2012, ICML.

[25]  Zhi-Hua Zhou,et al.  Multi-Label Learning with Weak Label , 2010, AAAI.

[26]  H. Zou,et al.  Regularization and variable selection via the elastic net , 2005 .

[27]  Ohad Shamir,et al.  Learning to classify with missing and corrupted features , 2008, ICML.

[28]  Quanquan Gu,et al.  Optimal Statistical and Computational Rates for One Bit Matrix Completion , 2016, AISTATS.