Improved Information Gain Estimates for Decision Tree Induction

Ensembles of classification and regression trees remain popular machine learning methods because they define flexible non-parametric models that predict well and are computationally efficient both during training and testing. During induction of decision trees one aims to find predicates that are maximally informative about the prediction target. To select good predicates most approaches estimate an information-theoretic scoring function, the information gain, both for classification and regression problems. We point out that the common estimation procedures are biased and show that by replacing them with improved estimators of the discrete and the differential entropy we can obtain better decision trees. In effect our modifications yield improved predictive performance and are simple to implement in any decision tree code.

[1]  Alfred O. Hero,et al.  Geodesic entropic graphs for dimension and entropy estimation in manifold learning , 2004, IEEE Transactions on Signal Processing.

[2]  Liam Paninski,et al.  Estimation of Entropy and Mutual Information , 2003, Neural Computation.

[3]  Ronald L. Rivest,et al.  Constructing Optimal Binary Decision Trees is NP-Complete , 1976, Inf. Process. Lett..

[4]  Philip J. Stone,et al.  Experiments in induction , 1966 .

[5]  Wei-Yin Loh,et al.  Classification and regression trees , 2011, WIREs Data Mining Knowl. Discov..

[6]  A. Antos,et al.  Convergence properties of functional estimates for discrete distributions , 2001 .

[7]  J. Ross Quinlan,et al.  Induction of Decision Trees , 1986, Machine Learning.

[8]  Sanjeev R. Kulkarni,et al.  Universal Estimation of Information Measures for Analog Sources , 2009, Found. Trends Commun. Inf. Theory.

[9]  Thomas Schürmann Bias analysis in entropy estimation , 2004 .

[10]  Jon Louis Bentley,et al.  An Algorithm for Finding Best Matches in Logarithmic Expected Time , 1977, TOMS.

[11]  Wray L. Buntine,et al.  A further comparison of splitting rules for decision-tree induction , 2004, Machine Learning.

[12]  P. Grassberger Entropy Estimates from Insufficient Samplings , 2003, physics/0307138.

[13]  D. Geman,et al.  Randomized Inquiries About Shape: An Application to Handwritten Digit Recognition. , 1994 .

[14]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[15]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[16]  Antonio Criminisi,et al.  Decision Forests: A Unified Framework for Classification, Regression, Density Estimation, Manifold Learning and Semi-Supervised Learning , 2012, Found. Trends Comput. Graph. Vis..

[17]  A. Buja,et al.  Loss Functions for Binary Class Probability Estimation and Classification: Structure and Applications , 2005 .

[18]  Wei Zhong Liu,et al.  The Importance of Attribute Selection Measures in Decision Tree Induction , 1994, Machine Learning.

[19]  L. Györfi,et al.  Nonparametric entropy estimation. An overview , 1997 .

[20]  D. V. Gokhale,et al.  Entropy expressions and their estimators for multivariate distributions , 1989, IEEE Trans. Inf. Theory.

[21]  Andreas Buja,et al.  Data mining criteria for tree-based regression and classification , 2001, KDD '01.

[22]  Ga Miller,et al.  Note on the bias of information estimates , 1955 .

[23]  P. Grassberger Finite sample corrections to entropy and dimension estimates , 1988 .

[24]  John Mingers,et al.  An Empirical Comparison of Selection Measures for Decision-Tree Induction , 1989, Machine Learning.

[25]  Alfred O. Hero,et al.  Asymptotic theory of greedy approximations to minimal k-point random graphs , 1999, IEEE Trans. Inf. Theory.

[26]  Robert A. Lordo,et al.  Nonparametric and Semiparametric Models , 2005, Technometrics.

[27]  M. N. Goria,et al.  A new class of random vector entropy estimators and its applications in testing statistical hypotheses , 2005 .

[28]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.