Semi-supervised Learning with Multimodal Perturbation

In this paper, a new co-training style semi-supervised algorithm is proposed, which employs Bagging based multimodal perturbation to label the unlabeled data. In detail, through perturbing the training data, input attributes and learning parameters together, the algorithm generates accurate but diversity k-nearest neighbor classifiers. These classifiers are refined using unlabeled examples which are labeled if the other classifiers agree on the labeling. Experimental results show that the semi-supervised algorithm could effectively improve the classification generalization by utilizing the unlabeled data.

[1]  Yan Zhou,et al.  Enhancing Supervised Learning with Unlabeled Data , 2000, ICML.

[2]  Tin Kam Ho,et al.  Nearest Neighbors in Random Subspaces , 1998, SSPR/SPR.

[3]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[4]  Avrim Blum,et al.  The Bottleneck , 2021, Monopsony Capitalism.

[5]  Zhi-Hua Zhou,et al.  Tri-training: exploiting unlabeled data using three classifiers , 2005, IEEE Transactions on Knowledge and Data Engineering.

[6]  Tony R. Martinez,et al.  Improved Heterogeneous Distance Functions , 1996, J. Artif. Intell. Res..

[7]  Yang Yu,et al.  Ensembling local learners ThroughMultimodal perturbation , 2005, IEEE Trans. Syst. Man Cybern. Part B.

[8]  Alexander Zien,et al.  Semi-Supervised Learning , 2006 .

[9]  Anders Krogh,et al.  Neural Network Ensembles, Cross Validation, and Active Learning , 1994, NIPS.

[10]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[11]  Xiaojin Zhu,et al.  --1 CONTENTS , 2006 .

[12]  Zhi-Hua Zhou,et al.  Improve Computer-Aided Diagnosis With Machine Learning Techniques Using Undiagnosed Samples , 2007, IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans.