Bayesian Network Classifiers: Model Approximation and Learning Complexity

Bayesian Networks are becoming increasingly popular in the machine learning community %, not just for inference but also for as a classification tool. When learning BN classifiers, due to lack of knowledge of true model, one makes a number of assumptions which lead to {\em modeling error}. The parameters of this model are further estimated from data, contributing to {\em estimation errors}. This paper analyzes both kinds of errors. We first bound the modeling error in terms of KL-distance between the true and the approximate distribution, and provide indications that ``often'''' this error is negligible. We then study the {\em learning complexity}, and bound the number of examples needed to minimize the estimation error in terms of the data {\em margin}.

[1]  Nir Friedman,et al.  Bayesian Network Classifiers , 1997, Machine Learning.

[2]  Linda Sellie,et al.  Toward efficient agnostic learning , 1992, COLT '92.

[3]  Dan Roth,et al.  Understanding Probabilistic Classifiers , 2001, ECML.

[4]  John Shawe-Taylor,et al.  A framework for structural risk minimisation , 1996, COLT '96.

[5]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[6]  Albert B Novikoff,et al.  ON CONVERGENCE PROOFS FOR PERCEPTRONS , 1963 .

[7]  Lawrence R. Rabiner,et al.  A tutorial on hidden Markov models and selected applications in speech recognition , 1989, Proc. IEEE.

[8]  Vladimir Pavlovic,et al.  Audio-visual speaker detection using dynamic Bayesian networks , 2000, Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580).

[9]  Dan Roth,et al.  Learning in Natural Language , 1999, IJCAI.

[10]  S. Kullback Probability Densities with Given Marginals , 1968 .