Double SVMBagging: A New Double Bagging with Support Vector Machine

In ensemble methods the aggregation of multiple unstable classifiers often leads to reduce the misclassification rates substantially in many applica- tions and benchmark classification problems. We pro- pose here a new ensemble, "Double SVMBagging", which is a variant of double bagging. In this ensem- ble method we used the support vector machine as the additional classifiers, built on the out-of-bag samples. The underlying base classifier is the decision tree. We used four kernel types; linear, polynomial, radial ba- sis and sigmoid kernels, expecting the new classifier perform in both linear and non-linear feature space. The major advantages of the proposed method is that, 1) it is compatible with the messy data structure, 2) the generation of support vectors in the first phase fa- cilitates the decision tree to classify the objects with higher confidence (accuracy), resulting in a significant error reduction in the second phase. We have applied the proposed method to a real case, the condition di- agnosis for the electric power apparatus; the feature variables are the maximum likelihood parameters in the generalized normal distribution, and weibull dis- tribution. These variables are composed from the partial discharge patterns of electromagnetic signals by the apparatus. We compare the performance of double SVMbagging with other well-known classifier ensemble methods in condition diagnosis; the double SVMbagging with the radial basis kernel performed better than other ensemble method and other kernels. We applied the double SVMbagging with radial basis kernel in 15 UCI benchmark datasets and compare it's accuracy with other ensemble methods e.g., Bag- ging, Adaboost, Random forest and Rotation Forest. The performance of this method demonstrates that this method can generate significantly lower predic- tion error than Rotation Forest and Adaboost more often than reverse. It performed much better than Bagging and Random Forest.

[1]  Pat Langley,et al.  Induction of One-Level Decision Trees , 1992, ML.

[2]  Yoav Freund,et al.  Experiments with a New Boosting Algorithm , 1996, ICML.

[3]  Bernhard Schölkopf,et al.  Nonlinear Component Analysis as a Kernel Eigenvalue Problem , 1998, Neural Computation.

[4]  Wei Zhang,et al.  Ensemble Implementations on Diversified Support Vector Machines , 2008, 2008 Fourth International Conference on Natural Computation.

[5]  De-Shuang Huang,et al.  Least squares support vector machine ensemble , 2004, 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541).

[6]  Wei-Yin Loh,et al.  Classification and regression trees , 2011, WIREs Data Mining Knowl. Discov..

[8]  Hideo Hirose,et al.  A New Double Bagging via the Support Vector Machine with Application to the Condition Diagnosis for the Electric Power Apparatus , 2009 .

[9]  H. Hirose,et al.  Diagnosis accuracy in electric power apparatus conditions using classification methods , 2010, IEEE Transactions on Dielectrics and Electrical Insulation.

[10]  Ye Li,et al.  Fault diagnosis based on support vector machine ensemble , 2005, 2005 International Conference on Machine Learning and Cybernetics.

[11]  Jason Weston,et al.  Support vector machines for multi-class pattern recognition , 1999, ESANN.

[12]  R. J. Patton,et al.  Artificial intelligence approaches to fault diagnosis , 1998 .

[13]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[14]  Leo Breiman,et al.  Statistical Modeling: The Two Cultures (with comments and a rejoinder by the author) , 2001 .

[15]  Javier Muguerza,et al.  A modular neural network approach to fault diagnosis , 1996, IEEE Trans. Neural Networks.

[16]  Torsten Hothorn,et al.  Double-Bagging: Combining Classifiers by Bootstrap Aggregation , 2002, Pattern Recognit..

[17]  J. Friedman Special Invited Paper-Additive logistic regression: A statistical view of boosting , 2000 .

[18]  Thorsten Joachims,et al.  Making large-scale support vector machine learning practical , 1999 .

[19]  John C. Platt,et al.  Fast training of support vector machines using sequential minimal optimization, advances in kernel methods , 1999 .

[20]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[21]  Johan A. K. Suykens,et al.  Benchmarking Least Squares Support Vector Machine Classifiers , 2004, Machine Learning.

[22]  Vladimir Cherkassky,et al.  The Nature Of Statistical Learning Theory , 1997, IEEE Trans. Neural Networks.

[23]  Giorgio Valentini,et al.  Low Bias Bagged Support Vector Machines , 2003, ICML.

[24]  H. Hirose,et al.  Electrical Insulation Diagnosing using a New Statistical Classification Method , 2006, 2006 IEEE 8th International Conference on Properties & applications of Dielectric Materials.

[25]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[26]  H. Hirose,et al.  Diagnosis of electric power apparatus using the decision tree method , 2008, IEEE Transactions on Dielectrics and Electrical Insulation.

[27]  Hao Wang,et al.  PSVM : Parallelizing Support Vector Machines on Distributed Computers , 2007 .

[28]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[29]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[30]  Juan José Rodríguez Diez,et al.  Rotation Forest: A New Classifier Ensemble Method , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[31]  Peter E. Hart,et al.  Nearest neighbor pattern classification , 1967, IEEE Trans. Inf. Theory.

[32]  Hsuan-Tien Lin,et al.  A note on Platt’s probabilistic outputs for support vector machines , 2007, Machine Learning.

[33]  John Platt,et al.  Probabilistic Outputs for Support vector Machines and Comparisons to Regularized Likelihood Methods , 1999 .

[34]  Kurt Hornik,et al.  Support Vector Machines in R , 2006 .

[35]  Hyun-Chul Kim,et al.  Constructing support vector machine ensemble , 2003, Pattern Recognit..

[36]  Edward Y. Chang,et al.  Parallelizing Support Vector Machines on Distributed Computers , 2007, NIPS.

[37]  Guangfu Ma,et al.  Support Vector Machines Ensemble Based on Fuzzy Integral for Classification , 2006, ISNN.

[38]  Chih-Jen Lin,et al.  Probability Estimates for Multi-class Classification by Pairwise Coupling , 2003, J. Mach. Learn. Res..

[39]  Kurt Hornik,et al.  The support vector machine under test , 2003, Neurocomputing.

[40]  Nello Cristianini,et al.  Advances in Kernel Methods - Support Vector Learning , 1999 .

[41]  Chih-Jen Lin,et al.  LIBLINEAR: A Library for Large Linear Classification , 2008, J. Mach. Learn. Res..

[42]  Sung-Bae Cho,et al.  Ensemble Approaches of Support Vector Machines for Multiclass Classification , 2007, PReMI.

[43]  K. Johana,et al.  Benchmarking Least Squares Support Vector Machine Classifiers , 2022 .

[44]  Leo Breiman,et al.  Statistical Modeling: The Two Cultures (with comments and a rejoinder by the author) , 2001, Statistical Science.