Recent work in Bayesian classifiers has shown that a better and more flexible representation of domain knowledge results in more accurate classifiers. We have recently examined a new type of Bayesian classifiers calledCase-Based Bayesian Network (CBBN) classifiers. The basic idea is to partition the training data into semantically sound clusters. A local BN classifier is then learned independently from each cluster. Such a flexible organization of domain knowledge can represent dependency assertions among attributes more accurately and more relevantly than possible in traditional Bayesian classifiers (i.e., BN and BMN classifiers), hence improving classification accuracy. RBMNs also provide a more flexible representation scheme than BNs and generalize BMNs. Briefly, a RBMN is a Decision Tree (DT) with component BNs at the leaves. In this paper, we further explore our CBBN classifiers by comparing them to RBMN classifiers. RBMNs partition the data using a DT induction algorithm. By contrast, CBBNs rely on a flexible strategy for clustering that handles outliers, therefore, allowing more freedom to search for the best way to cluster the data and represent the knowledge. Our experimental results show that CBBN classifiers perform significantly better than RBMN classifiers.
[1]
Weiru Liu,et al.
Learning belief networks from data: an information theory based approach
,
1997,
CIKM '97.
[2]
C. N. Liu,et al.
Approximating discrete probability distributions with dependence trees
,
1968,
IEEE Trans. Inf. Theory.
[3]
Nir Friedman,et al.
Bayesian Network Classifiers
,
1997,
Machine Learning.
[4]
Joshua Zhexue Huang,et al.
Extensions to the k-Means Algorithm for Clustering Large Data Sets with Categorical Values
,
1998,
Data Mining and Knowledge Discovery.
[5]
Russell Greiner,et al.
Learning Bayesian Belief Network Classifiers: Algorithms and System
,
2001,
Canadian Conference on AI.
[6]
Eugene Santos,et al.
Case-Based Bayesian Network Classifiers
,
2004,
FLAIRS.
[7]
Pedro Larrañaga,et al.
Learning Recursive Bayesian Multinets for Data Clustering by Means of Constructive Induction
,
2002,
Machine Learning.
[8]
Wai Lam,et al.
LEARNING BAYESIAN BELIEF NETWORKS: AN APPROACH BASED ON THE MDL PRINCIPLE
,
1994,
Comput. Intell..