A Rough Set Method for Co-training Algorithm

In recent years, semi-supervised learning has been a hot research topic in machine learning area. Different from traditional supervised learning which learns only from labeled data; semi-supervised learning makes use of both labeled and unlabeled data for learning purpose. Co-training is a popular semi-supervised learning algorithm which assumes that each example is represented by two or more redundantly sufficient sets of features (views) and additionally these views are independent given the class. To improve the performance and applicability of co-training, ensemble learning, such as bagging and random subspace has been used along with co-training. In this work, we propose to use the rough set based ensemble learning method with co-training algorithm (RSCO). Inherited the inherent characteristics of rough set, ensemble learning is expected to meet both the diversity and accuracy requirement. Finally experimental results on the UCI data sets demonstrate the promising performance of RSCO.

[1]  Rayid Ghani,et al.  Analyzing the effectiveness and applicability of co-training , 2000, CIKM '00.

[2]  S. Tsumoto,et al.  Rough set methods and applications: new developments in knowledge discovery in information systems , 2000 .

[3]  Greg Schohn,et al.  Less is More: Active Learning with Support Vector Machines , 2000, ICML.

[4]  Irena Koprinska,et al.  Co-training using RBF Nets and Different Feature Splits , 2006, The 2006 IEEE International Joint Conference on Neural Network Proceedings.

[5]  Avrim Blum,et al.  The Bottleneck , 2021, Monopsony Capitalism.

[6]  Daphne Koller,et al.  Support Vector Machine Active Learning with Applications to Text Classification , 2000, J. Mach. Learn. Res..

[7]  R. Słowiński Intelligent Decision Support: Handbook of Applications and Advances of the Rough Sets Theory , 1992 .

[8]  Sebastian Thrun,et al.  Text Classification from Labeled and Unlabeled Documents using EM , 2000, Machine Learning.

[9]  Günther Palm,et al.  Semi-supervised Learning of Tree-Structured RBF Networks Using Co-training , 2008, ICANN.

[10]  Krzysztof Krawiec,et al.  ROUGH SET REDUCTION OF ATTRIBUTES AND THEIR DOMAINS FOR NEURAL NETWORKS , 1995, Comput. Intell..

[11]  Thorsten Joachims,et al.  Transductive Inference for Text Classification using Support Vector Machines , 1999, ICML.

[12]  Zhi-Hua Zhou,et al.  Tri-training: exploiting unlabeled data using three classifiers , 2005, IEEE Transactions on Knowledge and Data Engineering.

[13]  Guoyin Wang,et al.  Rough reduction in algebra view and information view , 2003, Int. J. Intell. Syst..

[14]  Qingxiang Wu,et al.  Multiknowledge for decision making , 2005, Knowledge and Information Systems.

[15]  Tong Zhang,et al.  A Framework for Learning Predictive Structures from Multiple Tasks and Unlabeled Data , 2005, J. Mach. Learn. Res..

[16]  Yan Zhou,et al.  Enhancing Supervised Learning with Unlabeled Data , 2000, ICML.

[17]  Stan Matwin,et al.  Email classification with co-training , 2011, CASCON.

[18]  Sankar K. Pal,et al.  International Journal of Signal Processing , Image Processing and Pattern Recognition , 2008 .

[19]  Andrew McCallum,et al.  Employing EM and Pool-Based Active Learning for Text Classification , 1998, ICML.

[20]  Andrzej Skowron,et al.  Rough set methods in feature selection and recognition , 2003, Pattern Recognit. Lett..

[21]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[22]  Thorsten Joachims,et al.  Transductive Learning via Spectral Graph Partitioning , 2003, ICML.

[23]  Thomas G. Dietterich An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization , 2000, Machine Learning.

[24]  David Yarowsky,et al.  Unsupervised Word Sense Disambiguation Rivaling Supervised Methods , 1995, ACL.

[25]  Maria-Florina Balcan,et al.  A PAC-Style Model for Learning from Labeled and Unlabeled Data , 2005, COLT.

[26]  Siwei Luo,et al.  A random subspace method for co-training , 2008, 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence).

[27]  Nello Cristianini,et al.  Convex Methods for Transduction , 2003, NIPS.

[28]  Guoping Wang,et al.  Learning with progressive transductive Support Vector Machine , 2002, 2002 IEEE International Conference on Data Mining, 2002. Proceedings..

[29]  Mike Bauer,et al.  Proceedings of the 2001 conference of the Centre for Advanced Studies on Collaborative Research, November 5-7, 2001, Toronto, Ontario, Canada , 2001, CASCON.

[30]  Andrzej Skowron,et al.  The Discernibility Matrices and Functions in Information Systems , 1992, Intelligent Decision Support.

[31]  Francis K. H. Quek,et al.  Attribute bagging: improving accuracy of classifier ensembles by using random feature subsets , 2003, Pattern Recognit..

[32]  Ulf Brefeld,et al.  Semi-supervised learning for structured output variables , 2006, ICML.

[33]  Tin Kam Ho,et al.  The Random Subspace Method for Constructing Decision Forests , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[34]  Janusz Zalewski,et al.  Rough sets: Theoretical aspects of reasoning about data , 1996 .