Ameva: An autonomous discretization algorithm

This paper describes a new discretization algorithm, called Ameva, which is designed to work with supervised learning algorithms. Ameva maximizes a contingency coefficient based on Chi-square statistics and generates a potentially minimal number of discrete intervals. Its most important advantage, in contrast with several existing discretization algorithms, is that it does not need the user to indicate the number of intervals. We have compared Ameva with one of the most relevant discretization algorithms, CAIM. Tests performed comparing these two algorithms show that discrete attributes generated by the Ameva algorithm always have the lowest number of intervals, and even if the number of classes is high, the same computational complexity is maintained. A comparison between the Ameva and the genetic algorithm approaches has been also realized and there are very small differences between these iterative and combinatorial approaches, except when considering the execution time.

[1]  Lukasz A. Kurgan,et al.  CLIP4: Hybrid inductive machine learning algorithm that generates inequality rules , 2004, Inf. Sci..

[2]  Jason Catlett,et al.  On Changing Continuous Attributes into Ordered Discrete Attributes , 1991, EWSL.

[3]  Davide Anguita,et al.  Support vector machines for interval discriminant analysis , 2008, Neurocomputing.

[4]  William W. Cohen Fast Effective Rule Induction , 1995, ICML.

[5]  Chao-Ton Su,et al.  An Extended Chi2 Algorithm for Discretization of Real Value Attributes , 2005, IEEE Trans. Knowl. Data Eng..

[6]  Randy Kerber,et al.  ChiMerge: Discretization of Numeric Attributes , 1992, AAAI.

[7]  Francis Eng Hock Tay,et al.  A Modified Chi2 Algorithm for Discretization , 2002, IEEE Trans. Knowl. Data Eng..

[8]  Peter Clark,et al.  The CN2 Induction Algorithm , 1989, Machine Learning.

[9]  Robert C. Holte,et al.  Very Simple Classification Rules Perform Well on Most Commonly Used Datasets , 1993, Machine Learning.

[10]  David Zhang,et al.  Hand-Geometry Recognition Using Entropy-Based Discretization , 2007, IEEE Transactions on Information Forensics and Security.

[11]  Usama M. Fayyad,et al.  On the Handling of Continuous-Valued Attributes in Decision Tree Generation , 1992, Machine Learning.

[12]  D. J. Newman,et al.  UCI Repository of Machine Learning Database , 1998 .

[13]  Andrew McCallum,et al.  Dynamic conditional random fields: factorized probabilistic models for labeling and segmenting sequence data , 2004, J. Mach. Learn. Res..

[14]  Haym Hirsh,et al.  Converting numerical classification into text classification , 2003, Artif. Intell..

[15]  Andreu Català,et al.  Dual unification of bi-class support vector machine formulations , 2006, Pattern Recognit..

[16]  Lukasz A. Kurgan,et al.  CAIM discretization algorithm , 2004, IEEE Transactions on Knowledge and Data Engineering.

[17]  Andrew K. C. Wong,et al.  Class-Dependent Discretization for Inductive Learning from Continuous and Mixed-Mode Data , 1995, IEEE Trans. Pattern Anal. Mach. Intell..

[18]  Huan Liu,et al.  Feature Selection via Discretization , 1997, IEEE Trans. Knowl. Data Eng..

[19]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[20]  Duy-Dinh Le,et al.  Ent-Boost: Boosting Using Entropy Measure for Robust Object Detection , 2006, 18th International Conference on Pattern Recognition (ICPR'06).

[21]  Ron Kohavi,et al.  Supervised and Unsupervised Discretization of Continuous Features , 1995, ICML.

[22]  Marc Boullé,et al.  Khiops: A Statistical Discretization Method of Continuous Attributes , 2004, Machine Learning.

[23]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[24]  Kenneth A. Kaufman,et al.  Learning from Inconsistent and Noisy Data: The AQ18 Approach , 1999, ISMIS.