Improved C-Fuzzy Decision Trees

Pedrycz and Sosnowski proposed C-fuzzy decision trees based on information granulation. The tree grows gradually by using fuzzy C-means clustering algorithm to split the patterns in a selected node with the maximum heterogeneity into C corresponding children nodes. However, the distance function was only defined on the input difference between a pattern and a cluster center, causing difficulties in some cases. Besides, the output model of each leaf node represented by a constant restricts the representation capability about the data distribution in the node. We propose a more reasonable definition of the distance function by considering both the input and output differences with weighting factors. We also extend the output model of each leaf node to a local linear model and estimate the model parameters with a recursive SVD-based least squares estimator. Experimental results have shown that our improved version produces higher recognition rates and smaller mean square errors for classification and regression problems, respectively.

[1]  Shie-Jue Lee,et al.  A neuro-fuzzy system modeling with self-constructing rule generationand hybrid SVD-based learning , 2003, IEEE Trans. Fuzzy Syst..

[2]  Yinghua Lin,et al.  Using fuzzy partitions to create fuzzy systems from input-output data and set the initial weights in a fuzzy neural network , 1997, IEEE Trans. Fuzzy Syst..

[3]  Petra Perner,et al.  Data Mining - Concepts and Techniques , 2002, Künstliche Intell..

[4]  Witold Pedrycz,et al.  C-fuzzy decision trees , 2005, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[5]  J. Ross Quinlan,et al.  Induction of Decision Trees , 1986, Machine Learning.

[6]  Christopher J. Merz,et al.  UCI Repository of Machine Learning Databases , 1996 .