Model Trees for Classification of Hybrid Data Types

In the task of classification, most learning methods are suitable only for certain data types. For the hybrid dataset consists of nominal and numeric attributes, to apply the learning algorithms, some attributes must be transformed into the appropriate types. This procedure could damage the nature of dataset. We propose a model tree approach to integrate several characteristically different learning methods to solve the classification problem. We employ the decision tree as the classification framework and incorporate support vector machines into the tree construction process. This design removes the discretization procedure usually necessary for tree construction while decision tree induction itself can deal with nominal attributes which may not be handled well by e.g., SVM methods. Experiments show that our purposed method has better performance than that of other competing learning methods.

[1]  Yuh-Jye Lee,et al.  SSVM: A Smooth Support Vector Machine for Classification , 2001, Comput. Optim. Appl..

[2]  Eric R. Ziegel,et al.  The Elements of Statistical Learning , 2003, Technometrics.

[3]  Leo Breiman,et al.  Classification and Regression Trees , 1984 .

[4]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1997, EuroCOLT.

[5]  Helder Coelho Progress in Artificial Intelligence — IBERAMIA 98 , 2003, Lecture Notes in Computer Science.

[6]  James T. C. Teng,et al.  Multivariate decision trees using linear discriminants and tabu search , 2003, IEEE Trans. Syst. Man Cybern. Part A.

[7]  Tzuu-Hseng S. Li,et al.  Autonomous fuzzy parking control of a car-like mobile robot , 2003, IEEE Trans. Syst. Man Cybern. Part A.

[8]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[9]  David J. Spiegelhalter,et al.  Machine Learning, Neural and Statistical Classification , 2009 .

[10]  Simon Kasif,et al.  A System for Induction of Oblique Decision Trees , 1994, J. Artif. Intell. Res..

[11]  Simon Kasif,et al.  Induction of Oblique Decision Trees , 1993, IJCAI.

[12]  Aiko M. Hormann,et al.  Programs for Machine Learning. Part I , 1962, Inf. Control..

[13]  Usama M. Fayyad,et al.  Multi-Interval Discretization of Continuous-Valued Attributes for Classification Learning , 1993, IJCAI.

[14]  Ron Kohavi,et al.  Supervised and Unsupervised Discretization of Continuous Features , 1995, ICML.

[15]  Ron Kohavi,et al.  Error-Based and Entropy-Based Discretization of Continuous Features , 1996, KDD.

[16]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[17]  Jerome H. Friedman Multivariate adaptive regression splines (with discussion) , 1991 .

[18]  Carla E. Brodley,et al.  Linear Machine Decision Trees , 1991 .

[19]  K. Bennett,et al.  A support vector machine approach to decision trees , 1998, 1998 IEEE International Joint Conference on Neural Networks Proceedings. IEEE World Congress on Computational Intelligence (Cat. No.98CH36227).

[20]  J. Ross Quinlan,et al.  Bagging, Boosting, and C4.5 , 1996, AAAI/IAAI, Vol. 1.

[21]  J. Ross Quinlan,et al.  Improved Use of Continuous Attributes in C4.5 , 1996, J. Artif. Intell. Res..

[22]  Michael Schlosser,et al.  Non-Linear Decision Trees - NDT , 1996, ICML.

[23]  Luís Torgo,et al.  Dynamic Discretization of Continuous Attributes , 1998, IBERAMIA.

[24]  Simon Kasif,et al.  OC1: A Randomized Induction of Oblique Decision Trees , 1993, AAAI.

[25]  Carla E. Brodley,et al.  Multivariate decision trees , 2004, Machine Learning.

[26]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[27]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .