Parallel Approach for Ensemble Learning with Locally Coupled Neural Networks

Ensemble learning has gained considerable attention in different tasks including regression, classification and clustering. Adaboost and Bagging are two popular approaches used to train these models. The former provides accurate estimations in regression settings but is computationally expensive because of its inherently sequential structure, while the latter is less accurate but highly efficient. One of the drawbacks of the ensemble algorithms is the high computational cost of the training stage. To address this issue, we propose a parallel implementation of the Resampling Local Negative Correlation (RLNC) algorithm for training a neural network ensemble in order to acquire a competitive accuracy like that of Adaboost and an efficiency comparable to that of Bagging. We test our approach on both synthetic and real datasets from the UCI and Statlib repositories for the regression task. In particular, our fine-grained parallel approach allows us to achieve a satisfactory balance between accuracy and parallel efficiency.

[1]  Clay P. Breshears The Art of Concurrency - A Thread Monkey's Guide to Writing Parallel Applications , 2009 .

[2]  Harris Drucker,et al.  Improving Regressors using Boosting Techniques , 1997, ICML.

[3]  Robert E. Schapire,et al.  A Brief Introduction to Boosting , 1999, IJCAI.

[4]  Claudio Moraga,et al.  Local Negative Correlation with Resampling , 2006, IDEAL.

[5]  J. Freidman,et al.  Multivariate adaptive regression splines , 1991 .

[6]  William Gropp,et al.  MPICH2: A New Start for MPI Implementations , 2002, PVM/MPI.

[7]  Yves Grandvalet,et al.  Bagging Equalizes Influence , 2004, Machine Learning.

[8]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[9]  Thomas Hofmann,et al.  Map-Reduce for Machine Learning on Multicore , 2007 .

[10]  Dipti Verma,et al.  Data Mining: Next Generation Challenges and Future Directions , 2012 .

[11]  Pablo A. Estévez,et al.  A scalable parallel algorithm for training a hierarchical mixture of neural experts , 2002, Parallel Comput..

[12]  Jacek Gondzio,et al.  Hybrid MPI/OpenMP Parallel Linear Support Vector Machine Training , 2009 .

[13]  S. Vialle,et al.  AdaBoost Parallelization on PC Clusters with Virtual Shared Memory for Fast Feature Selection , 2007, 2007 IEEE International Conference on Signal Processing and Communications.

[14]  Gonzalo Martínez-Muñoz,et al.  Out-of-bag estimation of the optimal sample size in bagging , 2010, Pattern Recognit..

[15]  Durga L. Shrestha,et al.  Experiments with AdaBoost.RT, an Improved Boosting Scheme for Regression , 2006, Neural Computation.

[16]  Claudio Moraga,et al.  Ensemble Learning with Local Diversity , 2006, ICANN.

[17]  Christopher J. Merz,et al.  UCI Repository of Machine Learning Databases , 1996 .

[18]  Trevor Hastie,et al.  The Elements of Statistical Learning , 2001 .

[19]  Cesare Furlanello,et al.  Parallelizing AdaBoost by weights dynamics , 2007, Comput. Stat. Data Anal..

[20]  Gavin Brown,et al.  Diversity in neural network ensembles , 2004 .

[21]  Zoran Obradovic,et al.  Boosting Algorithms for Parallel and Distributed Learning , 2022 .

[22]  Yoav Freund,et al.  A Short Introduction to Boosting , 1999 .

[23]  Ana I. González Acuña An experimental comparison of three methods for constructing ensembles of decision trees: Bagging, Boosting, and Randomization , 2012 .

[24]  Fazilah Haron,et al.  A framework for grid-based neural networks , 2005, First International Conference on Distributed Frameworks for Multimedia Applications.

[25]  Thomas G. Dietterich An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization , 2000, Machine Learning.