Learning from Multiple Complementary Labels

Complementary-label learning is a new weakly-supervised learning framework that solves the problem where each training example is supplied with a complementary label, which only specifies one of the classes that the example does not belong to. Although a few works have demonstrated that an unbiased estimator of the original classification risk can be obtained from only complementarily labeled data, they are all restricted to the case where each example is associated with exactly one complementary label. It would be more promising to learn from multiple complementary labels simultaneously, as the supervision information would be richer if more complementary labels are provided. So far, whether there exists an unbiased risk estimator for learning from multiple complementary labels simultaneously is still unknown. In this paper, we will give an affirmative answer by deriving the first unbiased risk estimator for learning from multiple complementary labels. In addition, we further theoretically analyze the estimation error bound of our proposed approach, and show that the optimal parametric convergence rate is achieved. Finally, we experimentally demonstrate the effectiveness of the proposed approach.

[1]  N. Nathani,et al.  Foundations of Machine Learning , 2021, Introduction to AI Techniques for Renewable Energy Systems.

[2]  Kun Zhang,et al.  Generative-Discriminative Complementary Learning , 2019, AAAI.

[3]  Bo An,et al.  Partial Label Learning with Self-Guided Retraining , 2019, AAAI.

[4]  Christoph Meinel,et al.  Recurrent generative adversarial network for learning imbalanced medical image semantic segmentation , 2019, Multimedia Tools and Applications.

[5]  Masashi Sugiyama,et al.  Online Multiclass Classification Based on Prediction Margin for Partial Feedback , 2019, ArXiv.

[6]  Gang Niu,et al.  Complementary-Label Learning for Arbitrary Losses and Models , 2018, ICML.

[7]  Bo An,et al.  Leveraging Latent Label Distributions for Partial Label Learning , 2018, IJCAI.

[8]  Jian Yang,et al.  A Regularization Approach for Instance-Based Superset Label Learning , 2018, IEEE Transactions on Cybernetics.

[9]  Gang Niu,et al.  Classification from Pairwise Similarity and Unlabeled Data , 2018, ICML.

[10]  Dacheng Tao,et al.  Learning with Biased Complementary Labels , 2017, ECCV.

[11]  Gang Niu,et al.  Learning from Complementary Labels , 2017, NIPS.

[12]  Shin Ishii,et al.  Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[13]  Gang Niu,et al.  Positive-Unlabeled Learning with Non-Negative Risk Estimator , 2017, NIPS.

[14]  Aritra Ghosh,et al.  Robust Loss Functions under Label Noise for Deep Neural Networks , 2017, AAAI.

[15]  Cheng Soon Ong,et al.  Learning from Corrupted Binary Labels via Class-Probability Estimation , 2015, ICML.

[16]  Gang Niu,et al.  Convex Formulation for Learning from Positive and Unlabeled Data , 2015, ICML.

[17]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[18]  Gang Niu,et al.  Analysis of Learning from Positive and Unlabeled Data , 2014, NIPS.

[19]  Thomas G. Dietterich,et al.  Learnability of the Superset Label Learning Problem , 2014, ICML.

[20]  Nagarajan Natarajan,et al.  Learning with Noisy Labels , 2013, NIPS.

[21]  Gang Niu,et al.  Squared-loss Mutual Information Regularization: A Novel Information-theoretic Approach to Semi-supervised Learning , 2013, ICML.

[22]  Thomas G. Dietterich,et al.  A Conditional Multinomial Mixture Model for Superset Label Learning , 2012, NIPS.

[23]  Ben Taskar,et al.  Learning from Partial Labels , 2011, J. Mach. Learn. Res..

[24]  Xiaojin Zhu,et al.  Introduction to Semi-Supervised Learning , 2009, Synthesis Lectures on Artificial Intelligence and Machine Learning.

[25]  Rich Caruana,et al.  Classification with partial labels , 2008, KDD.

[26]  Charles Elkan,et al.  Learning classifiers from only positive and unlabeled data , 2008, KDD.

[27]  Shahar Mendelson,et al.  Lower Bounds for the Empirical Minimization Algorithm , 2008, IEEE Transactions on Information Theory.

[28]  Bernhard Schölkopf,et al.  Learning with Local and Global Consistency , 2003, NIPS.

[29]  Peter L. Bartlett,et al.  Rademacher and Gaussian Complexities: Risk Bounds and Structural Results , 2003, J. Mach. Learn. Res..

[30]  M. Talagrand,et al.  Probability in Banach Spaces: Isoperimetry and Processes , 1991 .

[31]  Colin McDiarmid,et al.  Surveys in Combinatorics, 1989: On the method of bounded differences , 1989 .