Parallelized Metaheuristic-Ensemble of Heterogeneous Feedforward Neural Networks for Regression Problems

A feedforward neural network ensemble trained through metaheuristic algorithms has been proposed by researchers to produce a group of optimal neural networks. This method, however, has proven to be very time-consuming during the optimization process. To overcome this limitation, we propose a metaheuristic-based learning algorithm for building an ensemble system, resulting in shorter training time. In our proposed method, a master–slave based metaheuristic algorithm is employed in the optimization process to produce a group of heterogeneous feedforward neural networks, in which the global search operations are executed on the master, and the tasks of objective evaluation are distributed to the slaves (workers). To reduce evaluation costs, the entire training dataset is randomly divided equally into several disjoint subsets. Each subset is randomly paired with another subset of the remainder and distributed to a worker for the objective evaluation. Following the optimization process, representative candidate solutions (individuals) from the entire population are selected to perform as the base components of the ensemble system. The performance of the proposed method has been compared with those of other state-of-the-art techniques in over 31 benchmark regression datasets taken from public repositories. The experimental results show that the proposed method not only reduces the computational time but also achieves significantly better prediction accuracy. Moreover, the proposed method achieved promising results in the application of a subset of the million song dataset, which identifies the release year of a song and predicts the buzz on Twitter.

[1]  Francisco Herrera,et al.  A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms , 2011, Swarm Evol. Comput..

[2]  Ponnuthurai N. Suganthan,et al.  Ensemble incremental learning Random Vector Functional Link network for short-term electric load forecasting , 2018, Knowl. Based Syst..

[3]  Xizhao Wang,et al.  A review on neural networks with random weights , 2018, Neurocomputing.

[4]  Mohammad Rasoul Narimani,et al.  A novel fuzzy adaptive configuration of particle swarm optimization to solve large-scale optimal reactive power dispatch , 2017, Appl. Soft Comput..

[5]  Ponnuthurai N. Suganthan,et al.  Heterogeneous comprehensive learning particle swarm optimization with enhanced exploration and exploitation , 2015, Swarm Evol. Comput..

[6]  Jianhua Yang,et al.  Genetic ensemble of extreme learning machine , 2014, Neurocomputing.

[7]  Alípio Mário Jorge,et al.  Ensemble approaches for regression: A survey , 2012, CSUR.

[8]  Gonzalo A. Ruz,et al.  A non-iterative method for pruning hidden neurons in neural networks with random weights , 2018, Appl. Soft Comput..

[9]  Y. H. Pao,et al.  Characteristics of the functional link net: a higher order delta rule net , 1988, IEEE 1988 International Conference on Neural Networks.

[10]  Dipayan Guha,et al.  Optimal tuning of 3 degree-of-freedom proportional-integral-derivative controller for hybrid distributed power system using dragonfly algorithm , 2018, Comput. Electr. Eng..

[11]  Dianhui Wang,et al.  Randomness in neural networks: an overview , 2017, WIREs Data Mining Knowl. Discov..

[12]  Tanachapong Wangchamhan,et al.  Efficient algorithms based on the k-means and Chaotic League Championship Algorithm for numeric, categorical, and mixed-type data clustering , 2017, Expert Syst. Appl..

[13]  Najdan Vukovic,et al.  A comprehensive experimental evaluation of orthogonal polynomial expanded random vector functional link neural networks for regression , 2017, Appl. Soft Comput..

[14]  Senén Barro,et al.  An extensive experimental survey of regression methods , 2019, Neural Networks.

[15]  G. Lewicki,et al.  Approximation by Superpositions of a Sigmoidal Function , 2003 .

[16]  Zhiwei Ni,et al.  Improved discrete artificial fish swarm algorithm combined with margin distance minimization for ensemble pruning , 2019, Comput. Ind. Eng..

[17]  R. Iman,et al.  Approximations of the critical region of the fbietkan statistic , 1980 .

[18]  Dejan J. Sobajic,et al.  Learning and generalization characteristics of the random vector Functional-link net , 1994, Neurocomputing.

[19]  Nasser L. Azad,et al.  Self-controlled bio-inspired extreme learning machines for scalable regression and classification: a comprehensive analysis with some recommendations , 2016, Artificial Intelligence Review.

[20]  Marc Parizeau,et al.  Analysis of a master-slave architecture for distributed evolutionary computations , 2006, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[21]  Dianhui Wang,et al.  Evolutionary extreme learning machine ensembles with size control , 2013, Neurocomputing.

[22]  Ling Tang,et al.  A non-iterative decomposition-ensemble learning paradigm using RVFL network for crude oil price forecasting , 2017, Appl. Soft Comput..

[23]  Sanjay Ghemawat,et al.  MapReduce: Simplified Data Processing on Large Clusters , 2004, OSDI.

[24]  Fei Han,et al.  An improved evolutionary extreme learning machine based on particle swarm optimization , 2013, Neurocomputing.

[25]  S. C. Choube,et al.  Optimal reactive power rescheduling based on EPSDE algorithm to enhance static voltage stability , 2014 .

[26]  Gavin Brown,et al.  Diversity in neural network ensembles , 2004 .

[27]  Xiaoyong Liu,et al.  Parameter optimization of support vector regression based on sine cosine algorithm , 2018, Expert Syst. Appl..

[28]  Robert P. W. Duin,et al.  Feedforward neural networks with random weights , 1992, Proceedings., 11th IAPR International Conference on Pattern Recognition. Vol.II. Conference B: Pattern Recognition Methodology and Systems.

[29]  Dianhui Wang,et al.  Distributed learning for Random Vector Functional-Link networks , 2015, Inf. Sci..

[30]  Tianyou Chai,et al.  An online learning neural network ensembles with random weights for regression of sequential data stream , 2017, Soft Comput..

[31]  Guoqiang Li,et al.  Fast learning network: a novel artificial neural network with a fast learning speed , 2013, Neural Computing and Applications.

[32]  John H. Holland,et al.  Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence , 1992 .

[33]  A. Kai Qin,et al.  Evolutionary extreme learning machine , 2005, Pattern Recognit..

[34]  Ratnadip Adhikari,et al.  A neural network based linear ensemble framework for time series forecasting , 2015, Neurocomputing.

[35]  Carlos Henggeler Antunes,et al.  Learning of a single-hidden layer feedforward neural network using an optimized extreme learning machine , 2014, Neurocomputing.

[36]  Jacek M. Zurada,et al.  Review and performance comparison of SVM- and ELM-based classifiers , 2014, Neurocomputing.

[37]  Pedro Melo-Pinto,et al.  Assessment of grapevine variety discrimination using stem hyperspectral data and AdaBoost of random weight neural networks , 2018, Appl. Soft Comput..

[38]  Andrew Lewis,et al.  The Whale Optimization Algorithm , 2016, Adv. Eng. Softw..

[39]  João Gama,et al.  Ensemble learning for data stream analysis: A survey , 2017, Inf. Fusion.

[40]  Behnam Mohammadi-Ivatloo,et al.  Short-term hydrothermal generation scheduling by a modified dynamic neighborhood learning based particle swarm optimization , 2015 .

[41]  Thomas G. Dietterich Multiple Classifier Systems , 2000, Lecture Notes in Computer Science.

[42]  Patrick Siarry,et al.  A survey on optimization metaheuristics , 2013, Inf. Sci..

[43]  Yilong Yin,et al.  An Improved Neural Network with Random Weights Using Backtracking Search Algorithm , 2015, Neural Processing Letters.

[44]  Ming Li,et al.  Insights into randomized algorithms for neural networks: Practical issues and common pitfalls , 2017, Inf. Sci..

[45]  François Kawala,et al.  Prédictions d'activité dans les réseaux sociaux en ligne , 2013 .

[46]  Hossam Faris,et al.  Optimizing connection weights in neural networks using the whale optimization algorithm , 2016, Soft Computing.

[47]  Iago A. Carvalho On the statistical evaluation of algorithmic's computational experimentation with infeasible solutions , 2019, Inf. Process. Lett..

[48]  Haixia Wang,et al.  Received Signal Strength Based Indoor Positioning Using a Random Vector Functional Link Network , 2018, IEEE Transactions on Industrial Informatics.

[49]  C. R. Rao,et al.  Generalized Inverse of Matrices and its Applications , 1972 .

[50]  P. N. Suganthan,et al.  A dynamic neighborhood learning based particle swarm optimizer for global numerical optimization , 2012, Inf. Sci..

[51]  Gavin Brown,et al.  Diversity and degrees of freedom in regression ensembles , 2018, Neurocomputing.

[52]  David H. Wolpert,et al.  No free lunch theorems for optimization , 1997, IEEE Trans. Evol. Comput..

[53]  Mehmet Fatih Tasgetiren,et al.  Differential evolution algorithm with ensemble of parameters and mutation strategies , 2011, Appl. Soft Comput..

[54]  En Zhu,et al.  DMP-ELMs: Data and model parallel extreme learning machines for large-scale learning tasks , 2018, Neurocomputing.

[55]  Y. Takefuji,et al.  Functional-link net computing: theory, system architecture, and functionalities , 1992, Computer.

[56]  Kurt Hornik,et al.  Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.

[57]  Adam P. Piotrowski,et al.  Performance of the air2stream model that relates air and stream water temperatures depends on the calibration method , 2018, Journal of Hydrology.

[58]  Wei Tang,et al.  Ensembling neural networks: Many could be better than all , 2002, Artif. Intell..

[59]  Bin Li,et al.  The extreme learning machine learning algorithm with tunable activation function , 2012, Neural Computing and Applications.

[60]  Hui Li,et al.  Research and development of neural network ensembles: a survey , 2018, Artificial Intelligence Review.

[61]  Yang Wang,et al.  Repairing the crossover rate in adaptive differential evolution , 2014, Appl. Soft Comput..

[62]  Stephen P. Boyd,et al.  Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers , 2011, Found. Trends Mach. Learn..

[63]  Adam P. Piotrowski,et al.  Swarm Intelligence and Evolutionary Algorithms: Performance versus speed , 2017, Inf. Sci..

[64]  Hossam Faris,et al.  Improving Extreme Learning Machine by Competitive Swarm Optimization and its application for medical diagnosis problems , 2018, Expert Syst. Appl..

[65]  Yu-Lin He,et al.  Random weight network-based fuzzy nonlinear regression for trapezoidal fuzzy number data , 2017, Appl. Soft Comput..

[66]  Bin Chen,et al.  Fast Learning Network with Parallel Layer Perceptrons , 2017, Neural Processing Letters.

[67]  Guang-Bin Huang,et al.  What are Extreme Learning Machines? Filling the Gap Between Frank Rosenblatt’s Dream and John von Neumann’s Puzzle , 2015, Cognitive Computation.

[68]  João P. P. Gomes,et al.  Ensemble of Efficient Minimal Learning Machines for Classification and Regression , 2017, Neural Processing Letters.

[69]  Fei Han,et al.  An improved learning algorithm for random neural networks based on particle swarm optimization and input-to-output sensitivity , 2019, Cognitive Systems Research.

[70]  Tianyou Chai,et al.  Combinatorial optimization of input features and learning parameters for decorrelated neural network ensemble-based soft measuring model , 2018, Neurocomputing.

[71]  Chang Feng,et al.  Meta-ELM: ELM with ELM hidden nodes , 2014, Neurocomputing.

[72]  Chee Kheong Siew,et al.  Extreme learning machine: Theory and applications , 2006, Neurocomputing.

[73]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[74]  Ali Husseinzadeh Kashan,et al.  League Championship Algorithm (LCA): An algorithm for global optimization inspired by sport championships , 2014, Appl. Soft Comput..

[75]  Václav Snásel,et al.  Metaheuristic design of feedforward neural networks: A review of two decades of research , 2017, Eng. Appl. Artif. Intell..

[76]  Seyedali Mirjalili,et al.  SCA: A Sine Cosine Algorithm for solving optimization problems , 2016, Knowl. Based Syst..

[77]  Nasser L. Azad,et al.  Optimally pruned extreme learning machine with ensemble of regularization techniques and negative correlation penalty applied to automotive engine coldstart hydrocarbon emission identification , 2014, Neurocomputing.

[78]  Kevin M. Passino,et al.  Stable Adaptive Control and Estimation for Nonlinear Systems , 2001 .

[79]  Hossam Faris,et al.  Metaheuristic-based extreme learning machines: a review of design formulations and applications , 2018, Int. J. Mach. Learn. Cybern..

[80]  Lars Kotthoff,et al.  The algorithm selection competitions 2015 and 2017 , 2018, Artif. Intell..

[81]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[82]  Adam P. Piotrowski,et al.  Differential Evolution algorithms applied to Neural Network training suffer from stagnation , 2014, Appl. Soft Comput..

[83]  P. N. Suganthan,et al.  A comprehensive evaluation of random vector functional link networks , 2016, Inf. Sci..

[84]  Dianhui Wang,et al.  Fast decorrelated neural network ensembles with random weights , 2014, Inf. Sci..

[85]  Guang-Bin Huang,et al.  An Insight into Extreme Learning Machines: Random Neurons, Random Features and Kernels , 2014, Cognitive Computation.

[86]  Xin Yao,et al.  Evolutionary ensembles with negative correlation learning , 2000, IEEE Trans. Evol. Comput..

[87]  R. Haupt,et al.  Antenna Design With a Mixed Integer Genetic Algorithm , 2007, IEEE Transactions on Antennas and Propagation.

[88]  Thierry Bertin-Mahieux,et al.  The Million Song Dataset , 2011, ISMIR.

[89]  Xin Yao,et al.  Diversity creation methods: a survey and categorisation , 2004, Inf. Fusion.

[90]  Yaochu Jin,et al.  A Competitive Swarm Optimizer for Large Scale Optimization , 2015, IEEE Transactions on Cybernetics.

[91]  Zhiping Lin,et al.  Self-Adaptive Evolutionary Extreme Learning Machine , 2012, Neural Processing Letters.

[92]  Peter Tiño,et al.  Managing Diversity in Regression Ensembles , 2005, J. Mach. Learn. Res..

[93]  Reza Olfati-Saber,et al.  Consensus and Cooperation in Networked Multi-Agent Systems , 2007, Proceedings of the IEEE.

[94]  Xin Yao,et al.  Making use of population information in evolutionary artificial neural networks , 1998, IEEE Trans. Syst. Man Cybern. Part B.

[95]  Seyedali Mirjalili,et al.  Dragonfly algorithm: a new meta-heuristic optimization technique for solving single-objective, discrete, and multi-objective problems , 2015, Neural Computing and Applications.

[96]  Amaury Lendasse,et al.  OP-ELM: Optimally Pruned Extreme Learning Machine , 2010, IEEE Transactions on Neural Networks.