A hybrid classification algorithm by subspace partitioning through semi-supervised decision tree

Among data mining techniques, the decision tree is one of the more widely used methods for building classification models in the real world because of its simplicity and ease of interpretation. However, the method has some drawbacks, including instability, the nonsmooth nature of the decision boundary, and the possibility of overfitting. To overcome these problems, several works have utilized the relative advantages of other classifiers, such as logistic regression, support vector machine, and neural networks, in combination with a decision tree, in hybrid models which avoid the drawbacks of other models. Some hybrid models have used decision trees to quickly and efficiently partition the input space, and many studies have proved the effectiveness of the hybrid methods. However, there is room for further improvement by considering the topological properties of a dataset, because typical decision trees split nodes based only on the target variable. The proposed semi-supervised decision tree splits internal nodes by utilizing both labels and the structural characteristics of data for subspace partitioning, to improve the accuracy of classifiers applied to terminal nodes in the hybrid models. Experimental results confirm the superiority of the proposed algorithm and demonstrate the detailed characteristics of the algorithm. HighlightsPropose the semi-supervised split criterion for decision trees.Combine the semi-supervised decision tree as subspace partitioning with other classifiers.Experiments on several datasets showed that the proposed method outperforms the existing ones.

[1]  G. G. Stokes "J." , 1890, The New Yale Book of Quotations.

[2]  Strother H. Walker,et al.  Estimation of the probability of an event as a function of several independent variables. , 1967, Biometrika.

[3]  David P. Doane,et al.  Aesthetic Frequency Classifications , 1976 .

[4]  J. Ross Quinlan,et al.  Improved Use of Continuous Attributes in C4.5 , 1996, J. Artif. Intell. Res..

[5]  D. Cox The Regression Analysis of Binary Sequences , 2017 .

[6]  K. Do,et al.  Combining non-parametric models with logistic regression: an application to motor vehicle injury data , 2000 .

[7]  Alex Alves Freitas,et al.  Automatic Design of Decision-Tree Induction Algorithms , 2015, SpringerBriefs in Computer Science.

[8]  Simon Haykin,et al.  Neural Networks: A Comprehensive Foundation , 1998 .

[9]  D. W. Scott On optimal and data based histograms , 1979 .

[10]  Herbert A. Sturges,et al.  The Choice of a Class Interval , 1926 .

[11]  Leo Breiman,et al.  Classification and Regression Trees , 1984 .

[12]  Ron Kohavi,et al.  Scaling Up the Accuracy of Naive-Bayes Classifiers: A Decision-Tree Hybrid , 1996, KDD.

[13]  D. Freedman,et al.  On the histogram as a density estimator:L2 theory , 1981 .

[14]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[15]  Ivor W. Tsang,et al.  Core Vector Machines: Fast SVM Training on Very Large Data Sets , 2005, J. Mach. Learn. Res..

[16]  Q. J. Wang DIRECT SAMPLE ESTIMATORS OF L MOMENTS , 1996 .

[17]  Richard P. Brent,et al.  Fast training algorithms for multilayer neural nets , 1991, IEEE Trans. Neural Networks.

[18]  J. Hosking L‐Moments: Analysis and Estimation of Distributions Using Linear Combinations of Order Statistics , 1990 .

[19]  L. Bottou,et al.  1 Support Vector Machine Solvers , 2007 .

[20]  Nello Cristianini,et al.  An Introduction to Support Vector Machines and Other Kernel-based Learning Methods , 2000 .

[21]  Wayne Ieee,et al.  Entropy Nets: From Decision Trees to Neural Networks , 1990 .

[22]  Sang Joon Kim,et al.  A Mathematical Theory of Communication , 2006 .

[23]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[24]  José Antonio Gómez-Ruiz,et al.  A combined neural network and decision trees model for prognosis of breast cancer relapse , 2003, Artif. Intell. Medicine.

[25]  Dan Steinberg,et al.  THE HYBRID CART-LOGIT MODEL IN CLASSIFICATION AND DATA MINING , 1998 .

[26]  J. Ross Quinlan,et al.  Induction of Decision Trees , 1986, Machine Learning.

[27]  Benjamin W. Heumann An Object-Based Classification of Mangroves Using a Hybrid Decision Tree - Support Vector Machine Approach , 2011, Remote. Sens..