Captured multi-label relations via joint deep supervised autoencoder

Abstract The mapping relations learning between instances and multiple labels should reflect the underlying joint probability distribution following by the data sets. The general solution of such problem is to assume that the samples are subject to a certain distribution, i.e. normal distribution, but this hypothesis cannot excavate the real underlying mapping relations hidden in the data sets. Meanwhile, it is not advisable to suppose that multiple labels are independent of each other. Therefore, we propose the deep supervised autoencoder as a generative model to learn the posterior conditional probability rather than assigning the specific distribution in advance. In this way, we propose the different joint augmented matrices of training instances X i and corresponding label sets Y i under the three multi-label relations assumptions as the inputs to learn the posterior probability distribution. Finally, the experiments under model assumptions are conducted on six data sets, and we also set different noise levels to verify whether the optimal hypothesis has the ability to handle the corrupted labels. Experiments on images, biology and music real-world data sets show that our method outperforms most of state-of-the-art multi-label classifiers.

[1]  Jianbin Qiu,et al.  A Combined Adaptive Neural Network and Nonlinear Model Predictive Control for Multirate Networked Industrial Process Control , 2016, IEEE Transactions on Neural Networks and Learning Systems.

[2]  Yurong Liu,et al.  A survey of deep neural network architectures and their applications , 2017, Neurocomputing.

[3]  Pengpeng Zhao,et al.  Active learning with label correlation exploration for multi-label image classification , 2017, IET Comput. Vis..

[4]  Shan Gao,et al.  Low-level structure feature extraction for image processing via stacked sparse denoising autoencoder , 2017, Neurocomputing.

[5]  Geoff Holmes,et al.  Classifier chains for multi-label classification , 2009, Machine Learning.

[6]  Jie Geng,et al.  Deep Supervised and Contractive Neural Network for SAR Image Classification , 2017, IEEE Transactions on Geoscience and Remote Sensing.

[7]  Hadi Sadoghi Yazdi,et al.  Label denoising based on Bayesian aggregation , 2017, Int. J. Mach. Learn. Cybern..

[8]  Abhijit Guha Roy,et al.  DASA: Domain adaptation in stacked autoencoders using systematic dropout , 2015, 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR).

[9]  Grigorios Tsoumakas,et al.  Multi-Label Classification: An Overview , 2007, Int. J. Data Warehous. Min..

[10]  Jiebo Luo,et al.  Weakly Semi-Supervised Deep Learning for Multi-Label Image Annotation , 2015, IEEE Transactions on Big Data.

[11]  Qingming Huang,et al.  Multi-label classification by exploiting local positive and negative pairwise label correlation , 2017, Neurocomputing.

[12]  Luca Martino,et al.  Multi-label methods for prediction with sequential data , 2015, Pattern Recognit..

[13]  Min-Ling Zhang,et al.  A Review on Multi-Label Learning Algorithms , 2014, IEEE Transactions on Knowledge and Data Engineering.

[14]  Y. Narahari,et al.  Multi-Label Classification from Multiple Noisy Sources Using Topic Models , 2017, Inf..

[15]  Yu-Chiang Frank Wang,et al.  Learning Deep Latent Spaces for Multi-Label Classification , 2017, ArXiv.

[16]  Neil D. Lawrence,et al.  Kernels for Vector-Valued Functions: a Review , 2011, Found. Trends Mach. Learn..

[17]  Eyke Hüllermeier,et al.  Dependent binary relevance models for multi-label classification , 2014, Pattern Recognit..

[18]  Michael S. Lew,et al.  Deep learning for visual understanding: A review , 2016, Neurocomputing.

[19]  Zhi-Hua Zhou,et al.  ML-KNN: A lazy learning approach to multi-label learning , 2007, Pattern Recognit..

[20]  Pawel Teisseyre,et al.  CCnet: Joint multi-label classification and feature selection using classifier chains and elastic net regularization , 2017, Neurocomputing.

[21]  Dacheng Tao,et al.  Classification with Noisy Labels by Importance Reweighting , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[22]  Jiebo Luo,et al.  Learning multi-label scene classification , 2004, Pattern Recognit..

[23]  Yoshua. Bengio,et al.  Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..

[24]  Qingming Huang,et al.  Joint Feature Selection and Classification for Multilabel Learning , 2018, IEEE Transactions on Cybernetics.

[25]  Zhiwu Lu,et al.  Learning from Weak and Noisy Labels for Semantic Segmentation , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[26]  Pascal Vincent,et al.  Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion , 2010, J. Mach. Learn. Res..

[27]  Hsuan-Tien Lin,et al.  Multilabel Classification with Principal Label Space Transformation , 2012, Neural Computation.

[28]  Frank Nielsen,et al.  Bregman Divergences and Surrogates for Learning , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[29]  Mahdieh Soleymani Baghshah,et al.  A probabilistic multi-label classifier with missing and noisy labels handling capability , 2017, Pattern Recognit. Lett..

[30]  Luca Martino,et al.  Efficient monte carlo methods for multi-dimensional learning with classifier chains , 2012, Pattern Recognit..

[31]  Yoshua Bengio,et al.  What regularized auto-encoders learn from the data-generating distribution , 2012, J. Mach. Learn. Res..

[32]  Fernando Pérez-Cruz,et al.  Deep Learning for Multi-label Classification , 2014, ArXiv.

[33]  Christoph Rensing,et al.  Multi-label Text Classification Using Semantic Features and Dimensionality Reduction with Autoencoders , 2017, LDK.

[34]  M. Verleysen,et al.  Classification in the Presence of Label Noise: A Survey , 2014, IEEE Transactions on Neural Networks and Learning Systems.

[35]  Renchu Guan,et al.  Multi-label Deep Learning for Gene Function Annotation in Cancer Pathways , 2018, Scientific Reports.

[36]  Chunxia Zhang,et al.  Discriminative Representation Learning with Supervised Auto-encoder , 2018, Neural Processing Letters.

[37]  Haibo He,et al.  Stacked Multilevel-Denoising Autoencoders: A New Representation Learning Approach for Wind Turbine Gearbox Fault Diagnosis , 2017, IEEE Transactions on Instrumentation and Measurement.

[38]  Heiko Paulheim,et al.  Local and global feature selection for multilabel classification with binary relevance , 2017, Artificial Intelligence Review.

[39]  Saso Dzeroski,et al.  An extensive experimental comparison of methods for multi-label learning , 2012, Pattern Recognit..

[40]  Jiwen Lu,et al.  Single Sample Face Recognition via Learning Deep Supervised Autoencoders , 2015, IEEE Transactions on Information Forensics and Security.