Competence-conscious associative classification

The classification performance of an associative classifier is strongly dependent on the statistic measure or metric that i s used to quantify the strength of the association between fea tures and classes (i.e., confidence, correlation etc.). Pre vious studies have shown that classifiers produced by differen t metrics may provide conflicting predictions, and that the be st metric to use is data-dependent and rarely known while designing the classifier. This uncertainty concerning the opt imal match between metrics and problems is a dilemma, and prevents associative classifiers to achieve their maximal p erformance. This dilemma is the focus of this paper. A possible solution to this dilemma is to learn the competence, expertise, or assertiveness of metrics. The basic idea is that each metric has a specific sub-domain for which it is most competent (i.e., it consistently produces more accurate classifiers than the ones produced by other metrics). Particularly, we investigate stacking-based me talearning methods, which use the training data to find the domain of competence of each metric. The meta-classifier describes the domains of competence (or areas of expertise) of each metric, enabling a more sensible use of these metrics so that competence-conscious classifiers can be produced (i.e., a metric is only used to produce classifiers for test instances that belong to its domain of competence). We conducted a systematic and comprehensive evaluation, usin g different datasets and evaluation measures, of classifiers produced by different metrics. The result is that, while no metric is always superior than all others, the selection of appropriate metrics according to their competence/expert ise (i.e., competence-conscious associative classifiers) see ms very effective, showing gains that range from 1.2% to 26.3% when compared to the baselines (SVMs and an existing ensemble method).

[1]  Yang Wang,et al.  Boosting an associative classifier , 2006, IEEE Transactions on Knowledge and Data Engineering.

[2]  João Gama,et al.  Cascade Generalization , 2000, Machine Learning.

[3]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[4]  Mykola Pechenizkiy,et al.  Dynamic Integration with Random Forests , 2006, ECML.

[5]  G. Yule,et al.  On the association of attributes in statistics, with examples from the material of the childhood society, &c , 1900, Proceedings of the Royal Society of London.

[6]  Alexander K. Seewald,et al.  Exploring the parameter state space of stacking , 2002, 2002 IEEE International Conference on Data Mining, 2002. Proceedings..

[7]  David H. Wolpert,et al.  Stacked generalization , 1992, Neural Networks.

[8]  Yang Wang,et al.  An Overview of Associative Classifiers , 2006, DMIN.

[9]  Mohammed J. Zaki,et al.  Multi-label Lazy Associative Classification , 2007, PKDD.

[10]  Mohammed J. Zaki,et al.  Multi-evidence, multi-criteria, lazy associative document classification , 2006, CIKM '06.

[11]  Jianyong Wang,et al.  HARMONY: Efficiently Mining the Best Rules for Classification , 2005, SDM.

[12]  Elena Baralis,et al.  Majority Classification by Means of Association Rules , 2003, PKDD.

[13]  Tomasz Imielinski,et al.  Mining association rules between sets of items in large databases , 1993, SIGMOD Conference.

[14]  Wynne Hsu,et al.  Integrating Classification and Association Rule Mining , 1998, KDD.

[15]  Jaideep Srivastava,et al.  Selecting the right interestingness measure for association patterns , 2002, KDD.

[16]  Peter A. Flach,et al.  Delegating classifiers , 2004, ICML.

[17]  Jiawei Han,et al.  Discriminative Frequent Pattern Analysis for Effective Classification , 2007, 2007 IEEE 23rd International Conference on Data Engineering.

[18]  Vikram Pudi,et al.  ACME: An Associative Classifier Based on Maximum Entropy Principle , 2005, ALT.

[19]  Wagner Meira,et al.  Learning to rank at query-time using association rules , 2008, SIGIR '08.

[20]  Philip S. Yu,et al.  Direct Discriminative Pattern Mining for Effective Classification , 2008, 2008 IEEE 24th International Conference on Data Engineering.

[21]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[22]  Gary Geunbae Lee,et al.  Efficient implementation of associative classifiers for document classification , 2007, Inf. Process. Manag..

[23]  Osmar R. Zaïane,et al.  Learning to Use a Learned Model: A Two-Stage Approach to Classification , 2006, Sixth International Conference on Data Mining (ICDM'06).

[24]  Howard J. Hamilton,et al.  Evaluation of Interestingness Measures for Ranking Discovered Knowledge , 2001, PAKDD.

[25]  Robert E. Schapire,et al.  A Brief Introduction to Boosting , 1999, IJCAI.

[26]  Mohammed J. Zaki,et al.  Lazy Associative Classification , 2006, Sixth International Conference on Data Mining (ICDM'06).

[27]  Alexander V. Nazin,et al.  Recursive Aggregation of Estimators by the Mirror Descent Algorithm with Averaging , 2005, Probl. Inf. Transm..

[28]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[29]  Philip S. Yu,et al.  Direct mining of discriminative and essential frequent patterns via model-based search tree , 2008, KDD.

[30]  Sotiris B. Kotsiantis,et al.  Bagged Voting Ensembles , 2004, AIMSA.

[31]  Thorsten Joachims,et al.  Training linear SVMs in linear time , 2006, KDD '06.

[32]  Ian H. Witten,et al.  Issues in Stacked Generalization , 2011, J. Artif. Intell. Res..

[33]  Keki B. Irani,et al.  Multi-interval discretization of continuos attributes as pre-processing for classi cation learning , 1993, IJCAI 1993.

[34]  Mykola Pechenizkiy,et al.  Diversity in Random Subspacing Ensembles , 2004, DaWaK.

[35]  Y. Freund Boosting a Weak Learning Algorithm by Majority to Be Published in Information and Computation , 1995 .

[36]  Shlomo Argamon,et al.  Arbitrating Among Competing Classifiers Using Learned Referees , 2001, Knowledge and Information Systems.

[37]  Peter A. Flach,et al.  Rule Evaluation Measures: A Unifying View , 1999, ILP.

[38]  Sanjay Chawla,et al.  CCCS: a top-down associative classifier for imbalanced class distribution , 2006, KDD '06.

[39]  Jian Pei,et al.  CMAR: accurate and efficient classification based on multiple class-association rules , 2001, Proceedings 2001 IEEE International Conference on Data Mining.