Constructing Rough Decision Forests

Decision forests are a type of classification paradigm which combines a collection of decision trees for a classification task, instead of depending on a single tree. Improvement of accuracy and stability is observed in experiments and applications. Some novel techniques to construct decision forests are proposed based on rough set reduction in this paper. As there are a lot of reducts for some data sets, a series of decision trees can be trained with different reducts. Three methods to select decision trees or reducts are presented, and decisions from selected trees are fused with the plurality voting rule. The experiments show that random selection is the worst solution in the proposed methods. It is also found that input diversity maximization doesn't guarantee output diversity maximization. Hence it cannot guarantee a good classification performance in practice. Genetic algorithm based selective rough decision forests consistently get good classification accuracies compared with a single tree trained by raw data as well as the other two forest constructing methods.

[1]  Thomas G. Dietterich An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization , 2000, Machine Learning.

[2]  Yen-Wei Chen,et al.  Random Independent Subspace for Face Recognition , 2004, KES.

[3]  Qinghua Hu,et al.  Combining multiple neural networks for classification based on rough set reduction , 2003, International Conference on Neural Networks and Signal Processing, 2003. Proceedings of the 2003.

[4]  Tin Kam Ho,et al.  The Random Subspace Method for Constructing Decision Forests , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[5]  Jiri Matas,et al.  On Combining Classifiers , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[6]  Wei Tang,et al.  Ensembling neural networks: Many could be better than all , 2002, Artif. Intell..

[7]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[8]  Andrzej Skowron,et al.  Rough set methods in feature selection and recognition , 2003, Pattern Recognit. Lett..

[9]  Raimondo Schettini,et al.  Automatic Classification Of Digital Photographs Based On Decision Forests , 2004, Int. J. Pattern Recognit. Artif. Intell..

[10]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1997, EuroCOLT.

[11]  Daren Yu,et al.  Entropies Of Fuzzy Indiscernibility Relation And Its Operations , 2005, Int. J. Uncertain. Fuzziness Knowl. Based Syst..

[12]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[13]  Qingxiang Wu,et al.  Multiknowledge for decision making , 2005, Knowledge and Information Systems.

[14]  Horst Bunke,et al.  Feature selection algorithms for the generation of multiple classifier systems and their application to handwritten word recognition , 2004 .

[15]  G DietterichThomas An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees , 2000 .

[16]  Lakhmi C. Jain,et al.  Knowledge-Based Intelligent Information and Engineering Systems , 2004, Lecture Notes in Computer Science.

[17]  Horst Bunke,et al.  Handwritten Word Recognition Using Classifier Ensembles Generated From Multiple Prototypes , 2004, Int. J. Pattern Recognit. Artif. Intell..

[18]  Weida Tong,et al.  Decision Forest: Combining the Predictions of Multiple Independent Decision Tree Models , 2003, J. Chem. Inf. Comput. Sci..