A Combinatorial Fusion Method for Feature Mining

This paper demonstrates how methods borrowed from information fusion can improve the performance of a classifier by constructing (“fusing”) new features that are combinations of existing numeric features. This work is an example of local pattern analysis and fusion because it identifies potentially useful patterns (i.e., feature combinations) from a single data source. In our work, we fuse features by mapping the numeric values for each feature to a rank and then averaging these ranks. The quality of the fused features is measured with respect to how well they classify minority-class examples, which makes this method especially effective for dealing with data sets that exhibit class imbalance. This paper evaluates our combinatorial feature fusion method on ten data sets, using three learning methods. The results indicate that our method can be quite effective in improving classifier performance, although it seems to improve the performance of some learning methods more than others. General Terms Algorithms, Performance, Experimentation

[1]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[2]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[3]  Haym Hirsh,et al.  Learning to Predict Rare Events in Event Sequences , 1998, KDD.

[4]  Yanhong Zhou,et al.  Feature Mining and Integration for Improving the Prediction Accuracy of Translation Initiation Sites in Eukaryotic mRNAs , 2006, 2006 Fifth International Conference on Grid and Cooperative Computing Workshops.

[5]  Peter A. Flach,et al.  The role of feature construction in inductive rule learning , 2000 .

[6]  Isabelle Guyon,et al.  An Introduction to Variable and Feature Selection , 2003, J. Mach. Learn. Res..

[7]  Andrew P. Bradley,et al.  The use of the area under the ROC curve in the evaluation of machine learning algorithms , 1997, Pattern Recognit..

[8]  Larry A. Rendell,et al.  Constructive Induction On Decision Trees , 1989, IJCAI.

[9]  D. Frank Hsu,et al.  Combinatorial Fusion Analysis: Methods and Practices of Combining Multiple Scoring Systems , 2006 .

[10]  Pat Langley,et al.  Selection of Relevant Features and Examples in Machine Learning , 1997, Artif. Intell..

[11]  Ryszard S. Michalski,et al.  Data-driven constructive induction in AQ17-PRE: A method and experiments , 1991, [Proceedings] Third International Conference on Tools for Artificial Intelligence - TAI 91.

[12]  Aiko M. Hormann,et al.  Programs for Machine Learning. Part I , 1962, Inf. Control..

[13]  Yoram Singer,et al.  Learning to Order Things , 1997, NIPS.

[14]  Stan Matwin,et al.  Feature Engineering for Text Classification , 1999, ICML.

[15]  Fernando E. B. Otero,et al.  Genetic Programming for Attribute Construction in Data Mining , 2002, EuroGP.

[16]  Gary M. Weiss Mining with rarity: a unifying framework , 2004, SKDD.

[17]  Ron Kohavi,et al.  The Case against Accuracy Estimation for Comparing Induction Algorithms , 1998, ICML.

[18]  D. Frank Hsu,et al.  Comparing Rank and Score Combination Methods for Data Fusion in Information Retrieval , 2005, Information Retrieval.

[19]  Salvatore J. Stolfo,et al.  Toward Scalable Learning with Non-Uniform Class and Cost Distributions: A Case Study in Credit Card Fraud Detection , 1998, KDD.

[20]  Ingrid Russell,et al.  An introduction to the WEKA data mining system , 2006, ITICSE '06.

[21]  Zijian Zheng,et al.  Constructing X-of-N Attributes for Decision Tree Learning , 2000, Machine Learning.

[22]  Jerzy W. Grzymala-Busse,et al.  An Approach to Imbalanced Data Sets Based on Changing Rule Strength , 2004, Rough-Neural Computing: Techniques for Computing with Words.