Rough reducts based SVM ensemble

Neural network ensemble has demonstrated many advantages over single neural networks in terms of generalization ability and parameters configuration. As compared to traditional bagging and boosting that are typical of generating individual network by horizontally partitioning training dataset, this paper proposes a new ensemble method called RRSE (rough reducts based SVM ensemble) that differs in its individual-generating technique. RRSE generates individual SVM (support vector machine) of ensemble by projection of training dataset on sufficient and necessary attribute sets (reducts). For the sake of structured/semi-structured data we are faced with under most circumstances, RRSE employs rough set theory to preserve the conditional attributes' dependency on decision attributes and generate all minimal reducts. Because of the distinct significance and classification ability among reducts, each individual networks of RRSE ensemble accordingly learns with wider variance and less computation complexity, thus achieving good ensemble generalization ability and learning efficiency.

[1]  Qiang Shen,et al.  Semantics-preserving dimensionality reduction: rough and fuzzy-rough-based approaches , 2004, IEEE Transactions on Knowledge and Data Engineering.

[2]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[3]  Salvatore Greco,et al.  Fuzzy Similarity Relation as a Basis for Rough Approximations , 1998, Rough Sets and Current Trends in Computing.

[4]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[5]  Francis K. H. Quek,et al.  Attribute bagging: improving accuracy of classifier ensembles by using random feature subsets , 2003, Pattern Recognit..

[6]  Lars Kai Hansen,et al.  Ensemble methods for handwritten digit recognition , 1992, Neural Networks for Signal Processing II Proceedings of the 1992 IEEE Workshop.

[7]  Didier Dubois,et al.  Putting Rough Sets and Fuzzy Sets Together , 1992, Intelligent Decision Support.

[8]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1997, EuroCOLT.

[9]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[10]  Yoav Freund,et al.  Boosting the margin: A new explanation for the effectiveness of voting methods , 1997, ICML.

[11]  Andrzej Skowron,et al.  The Discernibility Matrices and Functions in Information Systems , 1992, Intelligent Decision Support.

[12]  D. Vanderpooten Similarity Relation as a Basis for Rough Approximations , 1995 .

[13]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.