Toward naive Bayes with attribute value weighting

AbstractNaive Bayes makes an assumption regarding conditional independence, but this assumption rarely holds true in real-world applications, so numerous attempts have been made to relax this assumption. However, to the best of our knowledge, few studies have assigned different weights to different attribute values. In this study, we propose a new paradigm for a simple, efficient, and effective attribute value weighting approach called the correlation-based attribute value weighting approach (CAVW), which assigns a different weight to each attribute value by computing the difference between the attribute value-class correlation (relevance) and the average attribute value-attribute value intercorrelation (average redundancy). In CAVW, we use the information theoretic method with a strong theoretical background to assign different weights to different attribute values. Two different attribute value weighting measures called the mutual information (MI) measure and the Kullback–Leibler (KL) measure are employed, and thus two different versions are created, which we denote as CAVW-MI and CAVW-KL, respectively. According to extensive empirical studies based on a collection of 36 benchmark datasets from the University of California at Irvine repository, CAVW-MI and CAVW-KL both obtained more satisfactory experimental results compared with the naive Bayesian classifier and other four existing attribute weighting methods, and our methods also maintain the simplicity of the original naive Bayes model.

[1]  J. Ross Quinlan,et al.  C4.5: Programs for Machine Learning , 1992 .

[2]  อนิรุธ สืบสิงห์,et al.  Data Mining Practical Machine Learning Tools and Techniques , 2014 .

[3]  Jesús Alcalá-Fdez,et al.  KEEL Data-Mining Software Tool: Data Set Repository, Integration of Algorithms and Experimental Analysis Framework , 2011, J. Multiple Valued Log. Soft Comput..

[4]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[5]  S. García,et al.  An Extension on "Statistical Comparisons of Classifiers over Multiple Data Sets" for all Pairwise Comparisons , 2008 .

[6]  Diab M. Diab,et al.  Using differential evolution for fine tuning naïve Bayesian classifiers and its application for text classification , 2017, Appl. Soft Comput..

[7]  LIANGXIAO JIANG,et al.  Discriminatively Weighted Naive Bayes and its Application in Text Classification , 2012, Int. J. Artif. Intell. Tools.

[8]  Z. Cai,et al.  Improving Naive Bayes for Classification , 2010 .

[9]  Liangxiao Jiang,et al.  Weighted average of one-dependence estimators† , 2012, J. Exp. Theor. Artif. Intell..

[10]  Mong-Li Lee,et al.  SNNB: A Selective Neighborhood Based Naïve Bayes for Lazy Learning , 2002, PAKDD.

[11]  Hongwei Li,et al.  Bayesian network classifiers for probability-based metrics , 2013, J. Exp. Theor. Artif. Intell..

[12]  Dejing Dou,et al.  Calculating Feature Weights in Naive Bayes with Kullback-Leibler Measure , 2011, 2011 IEEE 11th International Conference on Data Mining.

[13]  Mark A. Hall,et al.  Correlation-based Feature Selection for Discrete and Numeric Class Machine Learning , 1999, ICML.

[14]  Geoffrey I. Webb,et al.  Not So Naive Bayes: Aggregating One-Dependence Estimators , 2005, Machine Learning.

[15]  Liangxiao Jiang,et al.  Not so greedy: Randomly Selected Naive Bayes , 2012, Expert Syst. Appl..

[16]  David Maxwell Chickering,et al.  Learning Bayesian Networks is , 1994 .

[17]  Geoffrey I. Webb,et al.  Alleviating naive Bayes attribute independence assumption by attribute weighting , 2013, J. Mach. Learn. Res..

[18]  Usama M. Fayyad,et al.  Multi-Interval Discretization of Continuous-Valued Attributes for Classification Learning , 1993, IJCAI.

[19]  Nir Friedman,et al.  Bayesian Network Classifiers , 1997, Machine Learning.

[20]  Hongwei Li,et al.  Naive Bayes for value difference metric , 2014, Frontiers of Computer Science.

[21]  Khalil el Hindi,et al.  Selectively Fine-Tuning Bayesian Network Learning Algorithm , 2016, Int. J. Pattern Recognit. Artif. Intell..

[22]  Peng Zhang,et al.  Toward value difference metric with attribute weighting , 2017, Knowledge and Information Systems.

[23]  Zhihua Cai,et al.  Attribute Weighting via Differential Evolution Algorithm for Attribute Weighted Naive Bayes (WNB) , 2011 .

[24]  Ron Kohavi,et al.  Scaling Up the Accuracy of Naive-Bayes Classifiers: A Decision-Tree Hybrid , 1996, KDD.

[25]  Liangxiao Jiang,et al.  A Novel Bayes Model: Hidden Naive Bayes , 2009, IEEE Transactions on Knowledge and Data Engineering.

[26]  Bernhard Pfahringer,et al.  Locally Weighted Naive Bayes , 2002, UAI.

[27]  Harry Zhang,et al.  Learning weighted naive Bayes with accurate ranking , 2004, Fourth IEEE International Conference on Data Mining (ICDM'04).

[28]  Liangxiao Jiang,et al.  Learning Tree Augmented Naive Bayes for Ranking , 2005, DASFAA.

[29]  Philip S. Yu,et al.  Top 10 algorithms in data mining , 2007, Knowledge and Information Systems.

[30]  Khalil el Hindi,et al.  Fine tuning the Naïve Bayesian learning algorithm , 2014, AI Commun..

[31]  Tommy W. S. Chow,et al.  Textual and Visual Content-Based Anti-Phishing: A Bayesian Approach , 2011, IEEE Transactions on Neural Networks.

[32]  Pat Langley,et al.  Induction of Selective Bayesian Classifiers , 1994, UAI.

[33]  Liangxiao Jiang,et al.  Instance Cloning Local Naive Bayes , 2005, Canadian Conference on AI.

[34]  Mark A. Hall,et al.  A decision tree-based attribute weighting filter for naive Bayes , 2006, Knowl. Based Syst..

[35]  Shasha Wang,et al.  Deep feature weighting for naive Bayes and its application to text classification , 2016, Eng. Appl. Artif. Intell..

[36]  Lukasz A. Kurgan,et al.  Knowledge discovery approach to automated cardiac SPECT diagnosis , 2001, Artif. Intell. Medicine.

[37]  Yoshua Bengio,et al.  Inference for the Generalization Error , 1999, Machine Learning.