AutoRWN: automatic construction and training of random weight networks using competitive swarm of agents

Random Weight Networks have been extensively used in many applications in the last decade because it has many strong features such as fast learning and good generalization performance. Most of the traditional training techniques for Random Weight Networks randomly select the connection weights and hidden biases and thus suffer from local optima stagnation and degraded convergence. The literature shows that stochastic population-based optimization techniques are well regarded and reliable alternative for Random Weight Networks optimization because of high local optima avoidance and flexibility. In addition, many practitioners and non-expert users find it difficult to set the other parameters of the network like the number of hidden neurons, the activation function, and the regularization factor. In this paper, an approach for training Random Weight Networks is proposed based on a recent variant of particle swarm optimization called competitive swarm optimization. Unlike most of Random Weight Networks training techniques, which are used to optimize only the input weights and hidden biases, the proposed approach will automatically tune the weights, biases, the number of hidden neurons, and regularization factor as well as the embedded activation function in the network, simultaneously. The goal is to help users to effectively identify a proper structure and hyperparameter values to their applications while obtaining reasonable prediction results. Twenty benchmark classification datasets are used to compare the proposed approach with different types of basic and hybrid Random Weight Network-based models. The experimental results on the benchmark datasets show that the reasonable classification results can be obtained by automatically tuning the hyperparameters using the proposed approach.

[1]  Khashayar Khorasani,et al.  Constructive feedforward neural networks using Hermite polynomial activation functions , 2005, IEEE Transactions on Neural Networks.

[2]  K. V. Arya,et al.  An effective gbest-guided gravitational search algorithm for real-parameter optimization and its application in training of feedforward neural networks , 2017, Knowl. Based Syst..

[3]  Yaochu Jin,et al.  A Competitive Swarm Optimizer for Large Scale Optimization , 2015, IEEE Transactions on Cybernetics.

[4]  Fei Han,et al.  An improved evolutionary extreme learning machine based on particle swarm optimization , 2013, Neurocomputing.

[5]  Hossam Faris,et al.  Optimizing the Learning Process of Feedforward Neural Networks Using Lightning Search Algorithm , 2016, Int. J. Artif. Intell. Tools.

[6]  John J. Grefenstette,et al.  Optimization of Control Parameters for Genetic Algorithms , 1986, IEEE Transactions on Systems, Man, and Cybernetics.

[7]  Hossam Faris,et al.  A competitive swarm optimizer with hybrid encoding for simultaneously optimizing the weights and structure of Extreme Learning Machines for classification problems , 2020, Int. J. Mach. Learn. Cybern..

[8]  David H. Wolpert,et al.  No free lunch theorems for optimization , 1997, IEEE Trans. Evol. Comput..

[9]  Andrew Lewis,et al.  S-shaped versus V-shaped transfer functions for binary Particle Swarm Optimization , 2013, Swarm Evol. Comput..

[10]  Chukiat Worasucheep,et al.  A Particle Swarm Optimization with stagnation detection and dispersion , 2008, 2008 IEEE Congress on Evolutionary Computation (IEEE World Congress on Computational Intelligence).

[11]  Hossam Faris,et al.  An intelligent system for spam detection and identification of the most relevant features based on evolutionary Random Weight Networks , 2019, Inf. Fusion.

[12]  Tianyou Chai,et al.  Multivariable dynamic modeling for molten iron quality using online sequential random vector functional-link networks with self-feedback connections , 2015, Inf. Sci..

[13]  Brian D. Ripley,et al.  Statistical aspects of neural networks , 1993 .

[14]  Jianzhong Wang,et al.  DOA Estimation of Excavation Devices with ELM and MUSIC-Based Hybrid Algorithm , 2017, Cognitive Computation.

[15]  Kuldip K. Paliwal,et al.  Bidirectional recurrent neural networks , 1997, IEEE Trans. Signal Process..

[16]  Xizhao Wang,et al.  A review on neural networks with random weights , 2018, Neurocomputing.

[17]  Zhiping Lin,et al.  Self-Adaptive Evolutionary Extreme Learning Machine , 2012, Neural Processing Letters.

[18]  Yoh-Han Pao,et al.  Stochastic choice of basis functions in adaptive function approximation and the functional-link net , 1995, IEEE Trans. Neural Networks.

[19]  R. Eberhart,et al.  Empirical study of particle swarm optimization , 1999, Proceedings of the 1999 Congress on Evolutionary Computation-CEC99 (Cat. No. 99TH8406).

[20]  Hossam Faris,et al.  Improving Extreme Learning Machine by Competitive Swarm Optimization and its application for medical diagnosis problems , 2018, Expert Syst. Appl..

[21]  Xuehua Zhao,et al.  An Efficient and Effective Automatic Recognition System for Online Recognition of Foreign Fibers in Cotton , 2016, IEEE Access.

[22]  Rainer Storn,et al.  Differential Evolution – A Simple and Efficient Heuristic for global Optimization over Continuous Spaces , 1997, J. Glob. Optim..

[23]  Yunpeng Ma,et al.  A Kind of Parameters Self-adjusting Extreme Learning Machine , 2016, Neural Processing Letters.

[24]  Shang-Liang Chen,et al.  Orthogonal least squares learning algorithm for radial basis function networks , 1991, IEEE Trans. Neural Networks.

[25]  Dae-Jong Lee,et al.  Parameter Optimization of Extreme Learning Machine Using Bacterial Foraging Algorithm , 2007 .

[26]  Wolfgang Frieß,et al.  Application of interpretable artificial neural networks to early monoclonal antibodies development. , 2019, European journal of pharmaceutics and biopharmaceutics : official journal of Arbeitsgemeinschaft fur Pharmazeutische Verfahrenstechnik e.V.

[27]  Gregory R. Madey,et al.  Heuristic and optimization approaches to extending the Kohonen self organizing algorithm , 1996 .

[28]  A. Kai Qin,et al.  Evolutionary extreme learning machine , 2005, Pattern Recognit..

[29]  Xizhao Wang,et al.  Voting-based instance selection from large data sets with MapReduce and random weight networks , 2016, Inf. Sci..

[30]  Dejan J. Sobajic,et al.  Learning and generalization characteristics of the random vector Functional-link net , 1994, Neurocomputing.

[31]  Hossam Faris,et al.  Improved monarch butterfly optimization for unconstrained global search and neural network training , 2018, Applied Intelligence.

[32]  Yongli Wang,et al.  Selfish herds optimization algorithm with orthogonal design and information update for training multi-layer perceptron neural network , 2018, Applied Intelligence.

[33]  Hossam Faris,et al.  Time-varying hierarchical chains of salps with random weight networks for feature selection , 2020, Expert Syst. Appl..

[34]  Bahram Gharabaghi,et al.  Extreme learning machine model for water network management , 2017, Neural Computing and Applications.

[35]  Yu-Lin He,et al.  Fuzzy nonlinear regression analysis using a random weight network , 2016, Inf. Sci..

[36]  A. E. Hoerl,et al.  Ridge regression: biased estimation for nonorthogonal problems , 2000 .

[37]  Antonia Azzini,et al.  Evolutionary ANNs: A state of the art survey , 2011, Intelligenza Artificiale.

[38]  Ying Lin,et al.  Particle Swarm Optimization With an Aging Leader and Challengers , 2013, IEEE Transactions on Evolutionary Computation.

[39]  Hossam Faris,et al.  Metaheuristic-based extreme learning machines: a review of design formulations and applications , 2018, Int. J. Mach. Learn. Cybern..

[40]  Hui Li,et al.  Evolutionary artificial neural networks: a review , 2011, Artificial Intelligence Review.

[41]  Yilmaz Kaya,et al.  Evaluation of texture features for automatic detecting butterfly species using extreme learning machine , 2014, J. Exp. Theor. Artif. Intell..

[42]  Daniel Svozil,et al.  Introduction to multi-layer feed-forward neural networks , 1997 .

[43]  De-Shuang Huang,et al.  A Neural Root Finder of Polynomials Based on Root Moments , 2004, Neural Computation.

[44]  Hossam Faris,et al.  Optimizing connection weights in neural networks using the whale optimization algorithm , 2016, Soft Computing.

[45]  D. R. Hush,et al.  Classification with neural networks: a performance analysis , 1989, IEEE 1989 International Conference on Systems Engineering.

[46]  Hossam Faris,et al.  Evolving Support Vector Machines using Whale Optimization Algorithm for spam profiles detection on online social networks in different lingual contexts , 2018, Knowl. Based Syst..

[47]  Le Zhang,et al.  An ensemble of decision trees with random vector functional link networks for multi-class classification , 2017, Appl. Soft Comput..

[48]  Hossam Faris,et al.  Salp Chain-Based Optimization of Support Vector Machines and Feature Weighting for Medical Diagnostic Information Systems , 2019, Algorithms for Intelligent Systems.