A Unified Framework for Decision Tree on Continuous Attributes

The standard algorithms of decision trees and their derived methods are usually constructed on the basis of the frequency information. However, they still suffer from a dilemma or multichotomous question for continuous attributes when two or more candidate cut points have the same or similar splitting performance with the optimal value, such as the maximal information gain ratio or the minimal Gini index. In this paper, we propose a unified framework model to deal with this question. We then design two algorithms based on Splitting Performance and the number of Expected Segments, called SPES1 and SPES2, which determine the optimal cut point, as follows. First, several candidate cut points are selected based on their splitting performances being the closest to the optimal. Second, we compute the number of expected segments for each candidate cut point. Finally, we combine these two measures by introducing a weighting factor $\alpha $ to determine the optimal one from several candidate cut points. To validate the effectiveness of our methods, we perform them on 25 benchmark datasets. The experimental results demonstrate that the classification accuracies of the proposed algorithms are superior to the current state-of-the-art methods in tackling the multichotomous question, about 5% in some cases. In particular, according to the proposed methods, the number of candidate cut points converges to a certain extent.

[1]  Rakesh Agrawal,et al.  SPRINT: A Scalable Parallel Classifier for Data Mining , 1996, VLDB.

[2]  Yingwei Jin,et al.  An effective discretization method for disposing high-dimensional data , 2014, Inf. Sci..

[3]  P. McCrone,et al.  Expanding the role of radiographers in reporting suspected lung cancer: A cost-effectiveness analysis using a decision tree model. , 2017, Radiography.

[4]  J. Ross Quinlan,et al.  Improved Use of Continuous Attributes in C4.5 , 1996, J. Artif. Intell. Res..

[5]  Karel Bartos,et al.  Learning data discretization via convex optimization , 2018, Machine Learning.

[6]  Yi-Hung Liu,et al.  Decision tree induction with a constrained number of leaf nodes , 2016, Applied Intelligence.

[7]  Steven Salzberg,et al.  Programs for Machine Learning , 2004 .

[8]  Habiba Drias,et al.  An intrusion detection and alert correlation approach based on revising probabilistic classifiers using expert knowledge , 2012, Applied Intelligence.

[9]  Khairul Anam,et al.  Evaluation of extreme learning machine for classification of individual and combined finger movements using electromyography on amputees and non-amputees , 2017, Neural Networks.

[10]  Ravi Kothari,et al.  A new node splitting measure for decision tree construction , 2010, Pattern Recognit..

[11]  P. Kasbekar,et al.  A Decision Tree Analysis of Diabetic Foot Amputation Risk in Indian Patients , 2017, Front. Endocrinol..

[12]  Jorma Rissanen,et al.  SLIQ: A Fast Scalable Classifier for Data Mining , 1996, EDBT.

[13]  Petros Drineas,et al.  Feature selection for linear SVM with provable guarantees , 2014, Pattern Recognit..

[14]  Jürgen Schmidhuber,et al.  Deep learning in neural networks: An overview , 2014, Neural Networks.

[15]  Joaquín Abellán,et al.  Analysis and extension of decision trees based on imprecise probabilities: Application on noisy data , 2014, Expert Syst. Appl..

[16]  Marco Loog,et al.  A benchmark and comparison of active learning for logistic regression , 2016, Pattern Recognit..

[17]  Maitreyee Dutta,et al.  Privacy preserving classification by using modified C4.5 , 2013, 2013 Sixth International Conference on Contemporary Computing (IC3).

[18]  N. Mookhambika,et al.  PRIVACY PRESERVING DECISION TREE LEARNING USING UNREALIZED DATA SETS , 2013 .

[19]  Keki B. Irani,et al.  Multi-interval discretization of continuos attributes as pre-processing for classi cation learning , 1993, IJCAI 1993.

[20]  Johannes Gehrke,et al.  BOAT—optimistic decision tree construction , 1999, SIGMOD '99.

[21]  Jiye Liang,et al.  Fusing Monotonic Decision Trees , 2015, IEEE Transactions on Knowledge and Data Engineering.

[22]  Ofir Ben-Assuli,et al.  Using Electronic Medical Records in Admission Decisions: A Cost Effectiveness Analysis , 2013, Decis. Sci..

[23]  Piotr Duda,et al.  New Splitting Criteria for Decision Trees in Stationary Data Streams , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[24]  Jiye Liang,et al.  A multi-view OVA model based on decision tree for multi-classification tasks , 2017, Knowl. Based Syst..

[25]  Leo Breiman,et al.  Classification and Regression Trees , 1984 .

[26]  Pierre Michel,et al.  Clustering nominal data using unsupervised binary decision trees: Comparisons with the state of the art methods , 2017, Pattern Recognit..

[27]  J. Ross Quinlan,et al.  Induction of Decision Trees , 1986, Machine Learning.

[28]  Xizhao Wang,et al.  Segment Based Decision Tree Induction With Continuous Valued Attributes , 2015, IEEE Transactions on Cybernetics.

[29]  Guido Bologna,et al.  Characterization of Symbolic Rules Embedded in Deep DIMLP Networks: A Challenge to Transparency of Deep Learning , 2017, J. Artif. Intell. Soft Comput. Res..

[30]  Francisco Herrera,et al.  Multivariate Discretization Based on Evolutionary Cut Points Selection for Classification , 2016, IEEE Transactions on Cybernetics.

[31]  Ye Chow Kuang,et al.  Sparse alternating decision tree , 2015, Pattern Recognit. Lett..

[32]  Kyoungok Kim,et al.  A hybrid classification algorithm by subspace partitioning through semi-supervised decision tree , 2016, Pattern Recognit..

[33]  Karim Jerbi,et al.  Learning machines and sleeping brains: Automatic sleep stage classification using decision-tree multi-class support vector machines , 2015, Journal of Neuroscience Methods.

[34]  Yong Wang,et al.  Online active learning of decision trees with evidential data , 2016, Pattern Recognit..

[35]  Deqin Yan,et al.  A new approach for discretizing continuous attributes in learning systems , 2014, Neurocomputing.

[36]  Francisco Herrera,et al.  A Survey of Discretization Techniques: Taxonomy and Empirical Analysis in Supervised Learning , 2013, IEEE Transactions on Knowledge and Data Engineering.

[37]  H. Kretser,et al.  Mobile decision-tree tool technology as a means to detect wildlife crimes and build enforcement networks , 2015 .

[38]  Khurram Shehzad,et al.  EDISC: A Class-Tailored Discretization Technique for Rule-Based Classification , 2012, IEEE Transactions on Knowledge and Data Engineering.

[39]  De Xu,et al.  Online State-Based Structured SVM Combined With Incremental PCA for Robust Visual Tracking , 2015, IEEE Transactions on Cybernetics.

[40]  Liming Yang,et al.  A sparse logistic regression framework by difference of convex functions programming , 2016, Applied Intelligence.

[41]  André Gustavo Maletzke,et al.  Exploring shapelet transformation for time series classification in decision trees , 2016, Knowl. Based Syst..

[42]  Vitaly Schetinin,et al.  Bayesian averaging over Decision Tree models for trauma severity scoring , 2018, Artif. Intell. Medicine.

[43]  Tapio Elomaa,et al.  General and Efficient Multisplitting of Numerical Attributes , 1999, Machine Learning.