A support vector machine approach to decision trees

Key ideas from statistical learning theory and support vector machines are generalized to decision trees. A support vector machine is used for each decision in the tree. The "optimal" decision tree is characterized, and both a primal and dual space formulation for constructing the tree are proposed. The result is a method for generating logically simple decision trees with multivariate linear, nonlinear or linear decisions. By varying the kernel function used, the decisions may consist of linear threshold units, polynomials, sigmoidal neural networks, or radial basis function networks. The preliminary results indicate that the method produces simple trees that generalize well with respect to other decision tree algorithms and single support vector machines.

[1]  Fred W. Glover,et al.  Tabu Search - Part I , 1989, INFORMS J. Comput..

[2]  Fred Glover,et al.  Tabu Search - Part II , 1989, INFORMS J. Comput..

[3]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[4]  Alberto Maria Segre,et al.  Programs for Machine Learning , 1994 .

[5]  Simon Kasif,et al.  A System for Induction of Oblique Decision Trees , 1994, J. Artif. Intell. Res..

[6]  Vladimir Vapnik,et al.  The Nature of Statistical Learning , 1995 .

[7]  Bernhard Schölkopf,et al.  Improving the Accuracy and Speed of Support Vector Machines , 1996, NIPS.

[8]  Bethany L. Nicholson,et al.  Mathematical Programs with Equilibrium Constraints , 2021, Pyomo — Optimization Modeling in Python.

[9]  TreesKristin P. Bennett,et al.  Optimal Decision Trees , 1996 .

[10]  Kristin P. Bennett,et al.  Hybrid extreme point tabu search , 1998, Eur. J. Oper. Res..