Semi-supervised Learning of Tree-Structured RBF Networks Using Co-training

Supervised learning requires a large amount of labeled data but the data labeling process can be expensive and time consuming. Co-training is a semi-supervised learning method that reduces the amount of required labeled data through exploiting the available unlabeled data in supervised learning to boost the accuracy. It assumes that the patterns are described by multiple independent feature sets and each feature set is sufficient for classification. On the other hand, most of the real-world pattern recognition tasks involve a large number of categories which may make the task difficult. The tree-structured approach is a multi-class decomposition strategy where a complex multi-class problem is decomposed into tree structured binary sub-problems. In this paper, we propose a framework that combines the tree-structured approach with Co-training. We show that our framework is especially useful for classification tasks involving a large number of classes and a small amount of labeled data where the tree-structured approach does not perform well by itself and when combined with Co-training, the unlabeled data boosts its accuracy.

[1]  Günther Palm,et al.  Comparison of Multiclass SVM Decomposition Schemes for Visual Object Recognition , 2005, DAGM-Symposium.

[2]  Rayid Ghani,et al.  Analyzing the effectiveness and applicability of co-training , 2000, CIKM '00.

[3]  Sameer A. Nene,et al.  Columbia Object Image Library (COIL100) , 1996 .

[4]  Friedhelm Schwenker,et al.  Three learning phases for radial-basis-function networks , 2001, Neural Networks.

[5]  Rebecca Fay,et al.  Feature selection and information fusion in hierarchical neural networks for iterative 3D-object recognition , 2007 .

[6]  Rayid Ghani,et al.  Combining labeled and unlabeled data for text classification with a large number of categories , 2001, Proceedings 2001 IEEE International Conference on Data Mining.

[7]  Günther Palm,et al.  Multi-View Forests of Tree-Structured Radial Basis Function Networks Based on Dempster-Shafer Evidence Theory , 2008, ESANN.

[8]  Tomas Hrycej Modular learning in neural networks - a modularized approach to neural network classification , 1992, Sixth-Generation computer technology series.

[9]  Arthur P. Dempster,et al.  A Generalization of Bayesian Inference , 1968, Classic Works of the Dempster-Shafer Theory of Belief Functions.

[10]  Glenn Shafer,et al.  A Mathematical Theory of Evidence , 2020, A Mathematical Theory of Evidence.

[11]  Avrim Blum,et al.  The Bottleneck , 2021, Monopsony Capitalism.

[12]  Matthias Seeger,et al.  Learning from Labeled and Unlabeled Data , 2010, Encyclopedia of Machine Learning.