Range Analysis and Applications to Root Causing

We propose a supervised learning algorithm whose aim is to derive features that explain the response variable better than the original features. Moreover, when there is a meaning for positive vs negative samples, our aim is to derive features that explain the positive samples, or subsets of positive samples that have the same root-cause. Each derived feature represents a single or multi-dimensional subspace of the feature space, where each dimension is specified as a feature-range pair for numeric features, and as a feature-level pair for categorical features. Unlike most Rule Learning and Subgroup Discovery algorithms, the response variable can be numeric, and our algorithm does not require a discretization of the response. The algorithm has been applied successfully to numerous real-life root-causing tasks in chip design, manufacturing, and validation, at Intel.

[1]  Pan Su,et al.  Feature Selection Ensemble , 2012, Turing-100.

[2]  Frank Puppe,et al.  Fast exhaustive subgroup discovery with numerical target concepts , 2016, Data Mining and Knowledge Discovery.

[3]  Aiko M. Hormann,et al.  Programs for Machine Learning. Part I , 1962, Inf. Control..

[4]  Isabelle Guyon,et al.  An Introduction to Variable and Feature Selection , 2003, J. Mach. Learn. Res..

[5]  S. Kotsiantis,et al.  Discretization Techniques: A recent survey , 2006 .

[6]  Florian Lemmerich,et al.  Fast Subgroup Discovery for Continuous Target Concepts , 2009, ISMIS.

[7]  Johannes Fürnkranz,et al.  Foundations of Rule Learning , 2012, Cognitive Technologies.

[8]  Amlan Chakrabarti,et al.  Feature Selection: A Practitioner View , 2014 .

[9]  Stefan Wrobel,et al.  An Algorithm for Multi-relational Discovery of Subgroups , 1997, PKDD.

[10]  Nicholas I. Fisher,et al.  Bump hunting in high-dimensional data , 1999, Stat. Comput..

[11]  Geoffrey I. Webb,et al.  Supervised Descriptive Rule Discovery: A Unifying Survey of Contrast Set, Emerging Pattern and Subgroup Mining , 2009, J. Mach. Learn. Res..

[12]  Peter Clark,et al.  The CN2 induction algorithm , 2004, Machine Learning.

[13]  Peter A. Flach,et al.  Subgroup Discovery with CN2-SD , 2004, J. Mach. Learn. Res..

[14]  Florian Lemmerich,et al.  pysubgroup: Easy-to-Use Subgroup Discovery in Python , 2018, ECML/PKDD.

[15]  J. Ross Quinlan,et al.  Induction of Decision Trees , 1986, Machine Learning.

[16]  Frank Puppe,et al.  Exploiting Background Knowledge for Knowledge-Intensive Subgroup Discovery , 2005, IJCAI.

[17]  C.J.H. Mann,et al.  Handbook of Data Mining and Knowledge Discovery , 2004 .

[18]  Yvan Saeys,et al.  Robust Feature Selection Using Ensemble Feature Selection Techniques , 2008, ECML/PKDD.

[19]  Randal S. Olson,et al.  PMLB: a large benchmark suite for machine learning evaluation and comparison , 2017, BioData Mining.

[20]  Benjamin Haibe-Kains,et al.  mRMRe: an R package for parallelized mRMR ensemble feature selection , 2013, Bioinform..

[21]  Chris Cornelis,et al.  Hybrid fuzzy-rough rule induction and feature selection , 2009, 2009 IEEE International Conference on Fuzzy Systems.

[22]  Willi Klösgen,et al.  Explora: A Multipattern and Multistrategy Discovery Assistant , 1996, Advances in Knowledge Discovery and Data Mining.

[23]  Ryszard S. Michalski,et al.  A theory and methodology of inductive learning , 1993 .

[24]  Chris H. Q. Ding,et al.  Minimum redundancy feature selection from microarray gene expression data , 2003, Computational Systems Bioinformatics. CSB2003. Proceedings of the 2003 IEEE Bioinformatics Conference. CSB2003.

[25]  Chris Cornelis,et al.  Applications of Fuzzy Rough Set Theory in Machine Learning: a Survey , 2015, Fundam. Informaticae.