A Novel AdaBoost Framework With Robust Threshold and Structural Optimization

The AdaBoost algorithm is a popular ensemble method that combines several weak learners to boost generalization performance. However, conventional AdaBoost.RT algorithms suffer from the limitation that the threshold value must be manually specified rather than chosen through a self-adaptive mechanism, which cannot guarantee a result in an optimal model for general cases. In this paper, we present a generic AdaBoost framework with robust threshold mechanism and structural optimization on regression problems. The error statistics of each weak learner on one given problem dataset is utilized to automate the choice of the optimal cut-off threshold value. In addition, a special single-layer neural network is employed to provide a second opportunity to further adjust the structure and strength the adaption capability of the AdaBoost regression model. Moreover, to consolidate the theoretical foundation of AdaBoost algorithms, we are the first to conduct a rigorous and comprehensive theoretical analysis on the proposed approach. We prove that the general bound on the empirical error with a fraction of training examples is always within a limited soft margin, which indicates that our novel algorithm can avoid over-fitting. We further analyze the bounds on the generalization error directly under probably approximately correct learning. The extensive experimental verifications on the UCI benchmarks have demonstrated that the performance of the proposed method is superior to other state-of-the-art ensemble and single learning algorithms. Furthermore, a real-world indoor positioning application has also revealed that the proposed method has higher positioning accuracy and faster speed.

[1]  Han Zou,et al.  Robust Extreme Learning Machine With its Application to Indoor Positioning , 2016, IEEE Transactions on Cybernetics.

[2]  A. Asuncion,et al.  UCI Machine Learning Repository, University of California, Irvine, School of Information and Computer Sciences , 2007 .

[3]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[4]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[5]  Tian Xia,et al.  A multi-class boosting method with direct optimization , 2014, KDD.

[6]  Harris Drucker,et al.  Improving Regressors using Boosting Techniques , 1997, ICML.

[7]  L. Breiman Arcing classifier (with discussion and a rejoinder by the author) , 1998 .

[8]  Wei Tang,et al.  Ensembling neural networks: Many could be better than all , 2002, Artif. Intell..

[9]  C. L. Philip Chen,et al.  A rapid supervised learning neural network for function interpolation and approximation , 1996, IEEE Trans. Neural Networks.

[10]  D.P. Solomatine,et al.  AdaBoost.RT: a boosting algorithm for regression problems , 2004, 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541).

[11]  David Haussler Quantifying the Inductive Bias in Concept Learning (Extended Abstract) , 1986, AAAI.

[12]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[13]  Hongming Zhou,et al.  Optimization method based extreme learning machine for classification , 2010, Neurocomputing.

[14]  Nathan Intrator,et al.  Boosting Regression Estimators , 1999, Neural Computation.

[15]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1997, EuroCOLT.

[16]  C. L. Philip Chen,et al.  A rapid learning and dynamic stepwise updating algorithm for flat neural networks and the application to time-series prediction , 1999, IEEE Trans. Syst. Man Cybern. Part B.

[17]  Chee Kheong Siew,et al.  Universal Approximation using Incremental Constructive Feedforward Networks with Random Hidden Nodes , 2006, IEEE Transactions on Neural Networks.

[18]  David Haussler,et al.  Probably Approximately Correct Learning , 2010, Encyclopedia of Machine Learning.

[19]  Chi-Chung Cheung,et al.  Magnified gradient function with deterministic weight modification in adaptive learning , 2004, IEEE Transactions on Neural Networks.

[20]  François Anctil,et al.  Forward Greedy ANN input selection in a stacked framework with Adaboost.RT - A streamflow forecasting case study exploiting radar rainfall estimates , 2012 .

[21]  Feng Jin,et al.  Image Restoration Based on an AdaBoost Algorithm , 2012 .

[22]  Johan A. K. Suykens,et al.  Least Squares Support Vector Machine Classifiers , 1999, Neural Processing Letters.

[23]  Cynthia Rudin,et al.  The Rate of Convergence of Adaboost , 2011, COLT.

[24]  Yoav Freund,et al.  Experiments with a New Boosting Algorithm , 1996, ICML.

[25]  Yimin Yang,et al.  Extreme Learning Machine With Subnetwork Hidden Nodes for Regression and Classification , 2016, IEEE Transactions on Cybernetics.

[26]  Zhixin Yang,et al.  A Robust AdaBoost.RT Based Ensemble Extreme Learning Machine , 2015 .

[27]  Lei Chen,et al.  RFID-enabled indoor positioning method for a real-time manufacturing execution system using OS-ELM , 2016, Neurocomputing.

[28]  Mehryar Mohri,et al.  Multi-Class Deep Boosting , 2014, NIPS.

[29]  Yoram Singer,et al.  Boosting with structural sparsity , 2009, ICML '09.

[30]  Yoav Freund,et al.  A Short Introduction to Boosting , 1999 .

[31]  Qiang Yang,et al.  Estimating Location Using Wi-Fi , 2008, IEEE Intelligent Systems.

[32]  Arman Savran,et al.  Regression-based intensity estimation of facial action units , 2012, Image Vis. Comput..

[33]  Homayoun Najjaran,et al.  Adaboost.MRT: Boosting regression for multivariate estimation , 2014, Artif. Intell. Res..

[34]  Alexander J. Smola,et al.  Support Vector Regression Machines , 1996, NIPS.

[35]  Hongming Zhou,et al.  Extreme Learning Machine for Regression and Multiclass Classification , 2012, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[36]  Zhi-Zhong Mao,et al.  An Ensemble ELM Based on Modified AdaBoost.RT Algorithm for Predicting the Temperature of Molten Steel in Ladle Furnace , 2010, IEEE Transactions on Automation Science and Engineering.

[37]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[38]  Yuan Yan Tang,et al.  ISABoost: A weak classifier inner structure adjusting based AdaBoost algorithm - ISABoost based application in scene categorization , 2013, Neurocomputing.

[39]  Lars Kai Hansen,et al.  Neural Network Ensembles , 1990, IEEE Trans. Pattern Anal. Mach. Intell..

[40]  Durga L. Shrestha,et al.  Experiments with AdaBoost.RT, an Improved Boosting Scheme for Regression , 2006, Neural Computation.

[41]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[42]  Yaonan Wang,et al.  Data Partition Learning With Multiple Extreme Learning Machines , 2015, IEEE Transactions on Cybernetics.

[43]  Samar K. Mukhopadhyay,et al.  A review of RFID technology and its managerial applications in different industries , 2012 .

[44]  Chunhua Shen,et al.  On the Dual Formulation of Boosting Algorithms , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[45]  O. Mangasarian,et al.  Robust linear programming discrimination of two linearly inseparable sets , 1992 .