Exploiting association and correlation rules parameters for learning Bayesian networks

In data mining, association and correlation rules are inferred from data in order to highlight statistical dependencies among attributes. The metrics defined for evaluating these rules can be exploited to score relationships between attributes in Bayesian network learning. In this paper, we propose two novel methods for learning Bayesian networks from data that are based on the K2 learning algorithm and that improve it by exploiting parameters normally defined for association and correlation rules. In particular, we propose the algorithms K2-Lift and K2-$X^{2}$, that exploit the lift metric and the $X^2$ metric respectively. We compare K2-Lift, K2-$X^{2}$ with K2 on artificial data and on three test Bayesian networks. The experiments show that both our algorithms improve K2 with respect to the quality of the learned network. Moreover, a comparison of K2-Lift and K2-$X^{2}$ with a genetic algorithm approach on two benchmark networks show superior results on one network and comparable results on the other.

[1]  David Maxwell Chickering,et al.  Learning Bayesian Networks: The Combination of Knowledge and Statistical Data , 1994, Machine Learning.

[2]  Moninder Singh,et al.  Construction of Bayesian network structures from data: A brief survey and an efficient algorithm , 1995, Int. J. Approx. Reason..

[3]  Shumeet Baluja,et al.  A Method for Integrating Genetic Search Based Function Optimization and Competitive Learning , 1994 .

[4]  Gregory F. Cooper,et al.  A Bayesian Method for the Induction of Probabilistic Networks from Data , 1992 .

[5]  Dan Geiger,et al.  An Entropy-based Learning Algorithm of Bayesian Conditional Trees , 1992, UAI.

[6]  J. A. Lozano,et al.  Estimation of Distribution Algorithms: A New Tool for Evolutionary Computation , 2001 .

[7]  Joe Suzuki,et al.  Learning Bayesian Belief Networks Based on the MDL Principle : An Efficient Algorithm Using the Branch and Bound Technique , 1999 .

[8]  Wai Lam,et al.  LEARNING BAYESIAN BELIEF NETWORKS: AN APPROACH BASED ON THE MDL PRINCIPLE , 1994, Comput. Intell..

[9]  Tomasz Imielinski,et al.  Mining association rules between sets of items in large databases , 1993, SIGMOD Conference.

[10]  D. Madigan,et al.  Model Selection and Accounting for Model Uncertainty in Graphical Models Using Occam's Window , 1994 .

[11]  Y. Yamashita,et al.  DNA microarray analysis of hematopoietic stem cell-like fractions from individuals with the M2 subtype of acute myeloid leukemia , 2003, Leukemia.

[12]  David Heckerman,et al.  A Tutorial on Learning with Bayesian Networks , 1999, Innovations in Bayesian Networks.

[13]  Paola Sebastiani,et al.  c ○ 2001 Kluwer Academic Publishers. Manufactured in The Netherlands. Robust Learning with Missing Data , 2022 .

[14]  Evelina Lamma,et al.  Bayesian Networks Learning for Gene Expression Datasets , 2005, IDA.

[15]  Pedro Larrañaga,et al.  Learning Bayesian networks in the space of structures by estimation of distribution algorithms , 2003, Int. J. Intell. Syst..

[16]  김현철 [서평]「Data Mining Techniques : For Marketing, Sales, and Customer Support」 , 1999 .

[17]  Rajeev Motwani,et al.  Beyond market baskets: generalizing association rules to correlations , 1997, SIGMOD '97.

[18]  Evelina Lamma,et al.  Exploiting Association and Correlation Rules - Parameters for Improving the K2 Algorithm , 2004, ECAI.

[19]  Sungzoon Cho,et al.  Constructing belief networks from realistic data , 1999 .

[20]  Pedro Larrañaga,et al.  Structure Learning of Bayesian Networks by Genetic Algorithms: A Performance Analysis of Control Parameters , 1996, IEEE Trans. Pattern Anal. Mach. Intell..

[21]  Heinz Mühlenbein,et al.  The Equation for Response to Selection and Its Use for Prediction , 1997, Evolutionary Computation.

[22]  David J. Spiegelhalter,et al.  Local computations with probabilities on graphical structures and their application to expert systems , 1990 .

[23]  Kevin B. Korb,et al.  Causal Discovery via MML , 1996, ICML.

[24]  Anna Goldenberg,et al.  Tractable learning of large Bayes net structures from sparse data , 2004, ICML.

[25]  Z. Estrov,et al.  The prognostic significance of p16INK4a/p14ARF locus deletion and MDM‐2 protein expression in adult acute myelogenous leukemia , 2000, Cancer.

[26]  Edward H. Herskovits,et al.  Computer-based probabilistic-network construction , 1992 .

[27]  Yong Liao,et al.  HER-2/neu induces p53 ubiquitination via Akt-mediated MDM2 phosphorylation , 2001, Nature Cell Biology.

[28]  Gregory Piatetsky-Shapiro,et al.  Discovery, Analysis, and Presentation of Strong Rules , 1991, Knowledge Discovery in Databases.

[29]  David A. Bell,et al.  Learning Bayesian networks from data: An information-theory based approach , 2002, Artif. Intell..

[30]  Silvia ACIDDepto BENEDICT : An Algorithm for Learning Probabilistic Belief Networks , 2007 .

[31]  Brent Boerlage Link Strength in Bayesian Networks , 1994 .

[32]  Gary Livingston,et al.  Evaluating the Causal Explanatory Value of Bayesian Network Structure Learning Algorithms , 2005 .

[33]  Richard E. Neapolitan,et al.  Learning Bayesian networks , 2007, KDD '07.

[34]  Michael J. A. Berry,et al.  Data mining techniques - for marketing, sales, and customer support , 1997, Wiley computer publishing.

[35]  Gregory F. Cooper,et al.  The ALARM Monitoring System: A Case Study with two Probabilistic Inference Techniques for Belief Networks , 1989, AIME.

[36]  Evelina Lamma,et al.  Improving the K2 Algorithm Using Association Rule Parameters , 2006 .

[37]  H. Mühlenbein,et al.  From Recombination of Genes to the Estimation of Distributions I. Binary Parameters , 1996, PPSN.

[38]  H. Akaike A new look at the statistical model identification , 1974 .

[39]  E. Estey,et al.  Plasma hepatocyte growth factor is a prognostic factor in patients with acute myeloid leukemia but not in patients with myelodysplastic syndrome , 2000, Leukemia.