Improved Behavior Knowledge Space Combination Method with Observational Learning

A new combination model, which incorporated observational learning with Behavior Knowledge Space (BKS) method, was proposed in this paper. Observational learning is a process with three steps: training, observing, and retraining. The proposed model generated simulated data by observational learning to extend the datasets, so it is effective in solving the small sample size problem that BKS suffers. Experimental investigations were performed on five datasets from the UCI repository. Bias-variance decomposition of the error indicates that observational learning algorithm can reduce both bias and variance. It is shown that, observational learning outperforms the individual base learner and majority voting when base learners are not capable enough for the given task, and classification performance can be improved further by repeat the "observing-retraining" process. Experiments also show that the combination model proposed in this work is superior to the basic BKS method and the BKS method with training dataset enlarged by injecting random noise.

[1]  Jiri Matas,et al.  On Combining Classifiers , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[2]  Peter L. Bartlett,et al.  Improved Generalization Through Explicit Optimization of Margins , 2000, Machine Learning.

[3]  Thomas G. Dietterich Multiple Classifier Systems , 2000, Lecture Notes in Computer Science.

[4]  Ching Y. Suen,et al.  A Method of Combining Multiple Experts for the Recognition of Unconstrained Handwritten Numerals , 1995, IEEE Trans. Pattern Anal. Mach. Intell..

[5]  Sungzoon Cho,et al.  Virtual sample generation using a population of networks , 2004, Neural Processing Letters.

[6]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[7]  Sungzoon Cho,et al.  Evolution of neural network training set through addition of virtual samples , 1996, Proceedings of IEEE International Conference on Evolutionary Computation.

[8]  Gian Luca Marcialis,et al.  An Experimental Comparison of Fixed and Trained Fusion Rules for Crisp Classifier Outputs , 2002, Multiple Classifier Systems.

[9]  Sungzoon Cho,et al.  Observational Learning Algorithm for an Ensemble of Neural Networks , 2002, Pattern Analysis & Applications.

[10]  Elie Bienenstock,et al.  Neural Networks and the Bias/Variance Dilemma , 1992, Neural Computation.

[11]  Yoav Freund,et al.  Boosting the margin: A new explanation for the effectiveness of voting methods , 1997, ICML.

[12]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[13]  Pedro M. Domingos A Unified Bias-Variance Decomposition for Zero-One and Squared Loss , 2000, AAAI/IAAI.

[14]  Noel E. Sharkey,et al.  The "Test and Select" Approach to Ensemble Combination , 2000, Multiple Classifier Systems.

[15]  Yoram Singer,et al.  Reducing Multiclass to Binary: A Unifying Approach for Margin Classifiers , 2000, J. Mach. Learn. Res..