An Efficient and Effective Case Classification Method Based On Slicing

One of the most important tasks that we have to face in real world applications is the task of classifying particular situations and /or events as belonging to a certain class. In order to solve the classification problem, accurate classifier systems or models must be built. Several computational intelligence methodologies have been applied to construct such a classifier from particular cases or data. This paper introduces a new classification method based on slicing techniques that was proposed for procedural programming languages. The paper also discusses two of common classification algorithms that are used either in data mining or in general AI. The algorithms are: Induction of Decision Tree Algorithm (ID3) and Base Learning Algorithm (C4.5). The paper also studies the comparison between the proposed method and the two selected classification algorithms using several domains.

[1]  David W. Aha,et al.  Towards a Better Understanding of Memory-based Reasoning Systems , 1994, ICML.

[2]  Morton Nadler,et al.  Pattern recognition engineering , 1993 .

[3]  J. Ross Quinlan,et al.  Induction of Decision Trees , 1986, Machine Learning.

[4]  Charles X. Ling,et al.  Toward Bayesian Classifiers with Accurate Probabilities , 2002, PAKDD.

[5]  David W. Binkley,et al.  Program slicing , 2008, 2008 Frontiers of Software Maintenance.

[6]  David W. Binkley,et al.  Interprocedural slicing using dependence graphs , 1988, SIGP.

[7]  Frank Tip,et al.  A survey of program slicing techniques , 1994, J. Program. Lang..

[8]  Hidehiko Tanaka,et al.  An Optimal Weighting Criterion of Case Indexing for Both Numeric and Symbolic Attributes , 1994 .

[9]  Pedro M. Domingos Rule Induction and Instance-Based Learning: A Unified Approach , 1995, IJCAI.

[10]  Thomas G. Dietterich,et al.  An Experimental Comparison of the Nearest-Neighbor and Nearest-Hyperrectangle Algorithms , 1995, Machine Learning.

[11]  Edwin Diday,et al.  A Recent Advance in Data Analysis: Clustering Objects into Classes Characterized by Conjunctive Concepts , 1981 .

[12]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[13]  Bernhard Pfahringer,et al.  Compression-Based Discretization of Continuous Attributes , 1995, ICML.

[14]  R. Tapia,et al.  Nonparametric Probability Density Estimation , 1978 .

[15]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[16]  Wamberto Weber Vasconcelos,et al.  Slicing knowledge-based systems: techniques and applications , 2000, Knowl. Based Syst..

[17]  Keith Brian Gallagher,et al.  Using Program Slicing in Software Maintenance , 1991, IEEE Trans. Software Eng..

[18]  Yoram Biberman,et al.  A Context Similarity Measure , 1994, ECML.

[19]  Andrew K. C. Wong,et al.  Class-Dependent Discretization for Inductive Learning from Continuous and Mixed-Mode Data , 1995, IEEE Trans. Pattern Anal. Mach. Intell..

[20]  Ron Kohavi,et al.  Supervised and Unsupervised Discretization of Continuous Features , 1995, ICML.

[21]  A. Tversky Features of Similarity , 1977 .

[22]  S. Salzberg,et al.  A weighted nearest neighbor algorithm for learning with symbolic features , 2004, Machine Learning.

[23]  S. Salzberg A nearest hyperrectangle learning method , 2004, Machine Learning.

[24]  Jason Catlett,et al.  On Changing Continuous Attributes into Ordered Discrete Attributes , 1991, EWSL.