Improving the K2 Algorithm Using Association Rule Parameters

Abstract A Bayesian network is an appropriate tool to work with the uncertainty that is typical of real–life applications.Bayesian network arcs represent statistical dependence between different variables and can be automatically elicited from database by Bayesian network learning algorithms such as K2. In the data mining field, association rules can also be interpreted as expressing statistical dependence relations. In this paper we present an extension of K2 called K2–rules that exploits a parameter normally defined in relation to association rules for learning Bayesian networks. We compare K2–rules with K2 and TPDA on the problems of learning four Bayesian networks. The experiments show that K2–rules improves both K2 and TPDA with respect to the quality of the learned network and K2 with respect to the execution time.

[1]  G. Schwarz Estimating the Dimension of a Model , 1978 .

[2]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[3]  David A. Bell,et al.  Learning Bayesian networks from data: An information-theory based approach , 2002, Artif. Intell..

[4]  H. Akaike A new look at the statistical model identification , 1974 .

[5]  Brent Boerlage Link Strength in Bayesian Networks , 1994 .

[6]  P. Spirtes,et al.  An Algorithm for Fast Recovery of Sparse Causal Graphs , 1991 .

[7]  David Heckerman,et al.  A Tutorial on Learning with Bayesian Networks , 1999, Innovations in Bayesian Networks.

[8]  Evelina Lamma,et al.  Improving the SLA Algorithm Using Association Rules , 2003, AI*IA.

[9]  David Maxwell Chickering,et al.  Learning Bayesian Networks: The Combination of Knowledge and Statistical Data , 1994, Machine Learning.

[10]  Gregory F. Cooper,et al.  A Bayesian Method for the Induction of Probabilistic Networks from Data , 1992 .

[11]  電子情報通信学会 IEICE transactions on communications electronics information and systems , 1991 .

[12]  Tomasz Imielinski,et al.  Mining association rules between sets of items in large databases , 1993, SIGMOD Conference.

[13]  Gregory F. Cooper,et al.  The ALARM Monitoring System: A Case Study with two Probabilistic Inference Techniques for Belief Networks , 1989, AIME.

[14]  Ramakrishnan Srikant,et al.  Fast algorithms for mining association rules , 1998, VLDB 1998.

[15]  David J. Spiegelhalter,et al.  Local computations with probabilities on graphical structures and their application to expert systems , 1990 .

[16]  Moninder Singh,et al.  Construction of Bayesian network structures from data: A brief survey and an efficient algorithm , 1995, Int. J. Approx. Reason..

[17]  Michael I. Jordan Learning in Graphical Models , 1999, NATO ASI Series.

[18]  Judea Pearl,et al.  A Computational Model for Causal and Diagnostic Reasoning in Inference Systems , 1983, IJCAI.

[19]  Edward H. Herskovits,et al.  Computer-based probabilistic-network construction , 1992 .

[20]  Joe Suzuki,et al.  Learning Bayesian Belief Networks Based on the MDL Principle : An Efficient Algorithm Using the Branch and Bound Technique , 1999 .

[21]  Gregory F. Cooper,et al.  A Bayesian method for the induction of probabilistic networks from data , 1992, Machine-mediated learning.

[22]  P. Spirtes,et al.  Causation, prediction, and search , 1993 .

[23]  D. Madigan,et al.  Model Selection and Accounting for Model Uncertainty in Graphical Models Using Occam's Window , 1994 .

[24]  C. N. Liu,et al.  Approximating discrete probability distributions with dependence trees , 1968, IEEE Trans. Inf. Theory.