A Study on Efficient Generation of Decision Trees Using Genetic Programming

For pattern recognition, the decision trees (DTs) are more efficient than neural networks (NNs) for two reasons. First, the computations in making decisions are simpler. Second, important features can be selected automatically during the design process. On the other hand, NNs are adaptable, and thus have the ability to learn in changing environment. Noting that there is a simple mapping from DT to NN, we can design a DT first, and then map it to an NN. By so doing, we can integrate the symbolic (DT) and the sub-symbolic (NN) approaches, and have advantages of both. For this purpose, we should design DTs which are as small as possible. In this paper, we continue our study on the evolutionary design of the decision trees based on genetic programming, and propose two new methods to reduce the tree sizes. The effectiveness of the new methods are tested through experiments with a character recognition problem.

[1]  William S. Meisel,et al.  A Partitioning Algorithm with Application in Pattern Classification and the Optimization of Decision Trees , 1973, IEEE Transactions on Computers.

[2]  Edward J. Delp,et al.  An Iterative Growing and Pruning Algorithm for Classification Tree Design , 1991, IEEE Trans. Pattern Anal. Mach. Intell..

[3]  Qiangfu Zhao,et al.  A study on evolutionary design of binary decision trees , 1999, Proceedings of the 1999 Congress on Evolutionary Computation-CEC99 (Cat. No. 99TH8406).

[4]  Ishwar K. Sethi,et al.  Entropy nets: from decision trees to neural networks , 1990, Proc. IEEE.

[5]  King-Sun Fu,et al.  A Nonparametric Partitioning Procedure for Pattern Classification , 1969, IEEE Transactions on Computers.

[6]  Hitoshi Iba,et al.  Genetic programming using a minimum description length principle , 1994 .

[7]  Ronald L. Rivest,et al.  Constructing Optimal Binary Decision Trees is NP-Complete , 1976, Inf. Process. Lett..