Regularized extreme learning machine for regression problems

Extreme learning machine (ELM) is a new learning algorithm for single-hidden layer feedforward networks (SLFNs) proposed by Huang et al. [1]. Its main advantage is the lower computational cost, which is especially relevant when dealing with many patterns defined in a high-dimensional space. This paper proposes an algorithm for pruning ELM networks by using regularized regression methods, thus obtaining a suitable number of the hidden nodes in the network architecture. Beginning from an initial large number of hidden nodes, irrelevant nodes are then pruned using ridge regression, elastic net and lasso methods; hence, the architectural design of ELM network can be automated. Empirical studies on several commonly used regression benchmark problems show that the proposed approach leads to compact networks that generate competitive results compared with the ELM algorithm.

[1]  K. S. Banerjee Generalized Inverse of Matrices and Its Applications , 1973 .

[2]  Trevor Hastie,et al.  The Elements of Statistical Learning , 2001 .

[3]  H. Akaike A new look at the statistical model identification , 1974 .

[4]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[5]  Zexuan Zhu,et al.  A fast pruned-extreme learning machine for classification problem , 2008, Neurocomputing.

[6]  Tao Xiong,et al.  A combined SVM and LDA approach for classification , 2005, Proceedings. 2005 IEEE International Joint Conference on Neural Networks, 2005..

[7]  Wenjian Wang,et al.  Optimal feed-forward neural networks based on the combination of constructing and pruning by genetic algorithms , 2002, Proceedings of the 2002 International Joint Conference on Neural Networks. IJCNN'02 (Cat. No.02CH37290).

[8]  Yi Zhao,et al.  A protein secondary structure prediction framework based on the Extreme Learning Machine , 2008, Neurocomputing.

[9]  J. Meulman,et al.  Prediction accuracy and stability of regression optimal scaling transformations , 1996 .

[10]  Chee Kheong Siew,et al.  Extreme learning machine: Theory and applications , 2006, Neurocomputing.

[11]  G. Schwarz Estimating the Dimension of a Model , 1978 .

[12]  Lila Ricci,et al.  Adjusted R-squared type measure for exponential dispersion models , 2010 .

[13]  C. R. Rao,et al.  Generalized Inverse of Matrices and its Applications , 1972 .

[14]  Trevor Hastie,et al.  Regularization Paths for Generalized Linear Models via Coordinate Descent. , 2010, Journal of statistical software.

[15]  Yong Yu,et al.  Sales forecasting using extreme learning machine with applications in fashion retailing , 2008, Decis. Support Syst..

[16]  N. S. Marimuthu,et al.  Intelligent approaches using support vector machine and extreme learning machine for transmission line protection , 2010, Neurocomputing.

[17]  Q. M. Jonathan Wu,et al.  Human action recognition using extreme learning machine based on visual vocabularies , 2010, Neurocomputing.

[18]  A. E. Hoerl,et al.  Ridge Regression: Applications to Nonorthogonal Problems , 1970 .

[19]  H. Zou,et al.  Regularization and variable selection via the elastic net , 2005 .