A non-iterative method for pruning hidden neurons in neural networks with random weights
暂无分享,去创建一个
[1] Juan de Oña,et al. Extracting the contribution of independent variables in neural network models: a new approach to handle instability , 2014, Neural Computing and Applications.
[2] Leonardo Ramos Rodrigues,et al. Building selective ensembles of Randomization Based Neural Networks with the successive projections algorithm , 2017, Appl. Soft Comput..
[3] Jingtao Yao,et al. Forecasting and Analysis of Marketing Data Using Neural Networks , 1998, J. Inf. Sci. Eng..
[4] Ivan Tyukin,et al. Feasibility of random basis function approximators for modeling and control , 2009, 2009 IEEE Control Applications, (CCA) & Intelligent Control, (ISIC).
[5] Le Zhang,et al. A survey of randomized algorithms for training neural networks , 2016, Inf. Sci..
[6] D. Lowe,et al. Adaptive radial basis function nonlinearities, and the problem of generalisation , 1989 .
[7] A. T. C. Goh,et al. Back-propagation neural networks for modeling complex systems , 1995, Artif. Intell. Eng..
[8] Shaoning Pang,et al. Incremental Learning of Chunk Data for Online Pattern Classification Systems , 2008, IEEE Transactions on Neural Networks.
[9] Robert P. W. Duin,et al. Feedforward neural networks with random weights , 1992, Proceedings., 11th IAPR International Conference on Pattern Recognition. Vol.II. Conference B: Pattern Recognition Methodology and Systems.
[10] M. Friedman. The Use of Ranks to Avoid the Assumption of Normality Implicit in the Analysis of Variance , 1937 .
[11] Dejan J. Sobajic,et al. Learning and generalization characteristics of the random vector Functional-link net , 1994, Neurocomputing.
[12] Janez Demsar,et al. Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..
[13] I. Dimopoulos,et al. Role of some environmental variables in trout abundance models using neural networks , 1996 .
[14] G. David Garson,et al. Interpreting neural-network connection weights , 1991 .
[15] R. Polikar,et al. Multiple Classifiers Based Incremental Learning Algorithm for Learning in Nonstationary Environments , 2007, 2007 International Conference on Machine Learning and Cybernetics.
[16] Narasimhan Sundararajan,et al. A generalized growing and pruning RBF (GGAP-RBF) neural network for function approximation , 2005, IEEE Transactions on Neural Networks.
[17] Yann LeCun,et al. Regularization of Neural Networks using DropConnect , 2013, ICML.
[18] Kurt Hornik,et al. Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.
[19] Alex Pappachen James,et al. Hierarchical Temporal Memory Features with Memristor Logic Circuits for Pattern Recognition , 2018, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.
[20] Ming Li,et al. Insights into randomized algorithms for neural networks: Practical issues and common pitfalls , 2017, Inf. Sci..
[21] Allan Pinkus,et al. Multilayer Feedforward Networks with a Non-Polynomial Activation Function Can Approximate Any Function , 1991, Neural Networks.
[22] Haixia Wang,et al. Received Signal Strength Based Indoor Positioning Using a Random Vector Functional Link Network , 2018, IEEE Transactions on Industrial Informatics.
[23] Gonzalo A. Ruz,et al. Extreme learning machine with a deterministic assignment of hidden weights in two parallel layers , 2017, Neurocomputing.
[24] Jooyoung Park,et al. Universal Approximation Using Radial-Basis-Function Networks , 1991, Neural Computation.
[25] Zexuan Zhu,et al. A fast pruned-extreme learning machine for classification problem , 2008, Neurocomputing.
[26] Chong-Yu Xu,et al. Relative Importance Analysis of a Refined Multi-parameter Phosphorus Index Employed in a Strongly Agriculturally Influenced Watershed , 2015, Water, Air, & Soil Pollution.
[27] Ling Tang,et al. A non-iterative decomposition-ensemble learning paradigm using RVFL network for crude oil price forecasting , 2017, Appl. Soft Comput..
[28] Najdan Vukovic,et al. A comprehensive experimental evaluation of orthogonal polynomial expanded random vector functional link neural networks for regression , 2017, Appl. Soft Comput..
[29] Corinna Cortes,et al. Support-Vector Networks , 1995, Machine Learning.
[30] Ponnuthurai N. Suganthan,et al. Random vector functional link network for short-term electricity load demand forecasting , 2016, Inf. Sci..
[31] Kurt Hornik,et al. Multilayer feedforward networks are universal approximators , 1989, Neural Networks.
[32] P. N. Suganthan,et al. A comprehensive evaluation of random vector functional link networks , 2016, Inf. Sci..
[33] Yoh-Han Pao,et al. Stochastic choice of basis functions in adaptive function approximation and the functional-link net , 1995, IEEE Trans. Neural Networks.
[34] Amaury Lendasse,et al. OP-ELM: Optimally Pruned Extreme Learning Machine , 2010, IEEE Transactions on Neural Networks.
[35] Osamu Fujita. Trial-and-error correlation learning , 1993, IEEE Trans. Neural Networks.
[36] Ivan Tyukin,et al. Approximation with random bases: Pro et Contra , 2015, Inf. Sci..
[37] D. Broomhead,et al. Radial Basis Functions, Multi-Variable Functional Interpolation and Adaptive Networks , 1988 .
[38] Ajalmar R. da Rocha Neto,et al. A new pruning method for extreme learning machines via genetic algorithms , 2016, Appl. Soft Comput..
[39] Gonzalo A. Ruz,et al. An Empirical Study of the Hidden Matrix Rank for Neural Networks with Random Weights , 2017, 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA).
[40] C. L. Philip Chen,et al. A rapid supervised learning neural network for function interpolation and approximation , 1996, IEEE Trans. Neural Networks.
[41] Simon Haykin,et al. Neural Networks: A Comprehensive Foundation , 1998 .
[42] K. S. Banerjee. Generalized Inverse of Matrices and Its Applications , 1973 .
[43] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[44] Yannis Dimopoulos,et al. Use of some sensitivity criteria for choosing networks with good generalization ability , 1995, Neural Processing Letters.
[45] M. Friedman. A Comparison of Alternative Tests of Significance for the Problem of $m$ Rankings , 1940 .
[46] Guang-Bin Huang,et al. Extreme learning machine: a new learning scheme of feedforward neural networks , 2004, 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541).
[47] M. Gevrey,et al. Review and comparison of methods to study the contribution of variables in artificial neural network models , 2003 .
[48] Narasimhan Sundararajan,et al. A Fast and Accurate Online Sequential Learning Algorithm for Feedforward Networks , 2006, IEEE Transactions on Neural Networks.
[49] Katarzyna Pentoź. The methods of extracting the contribution of variables in artificial neural network models - Comparison of inherent instability , 2016 .
[50] Yuwei Cui,et al. Continuous Online Sequence Learning with an Unsupervised Neural Network Model , 2015, Neural Computation.