Extreme Learning Machine with Fuzzy Activation Function
暂无分享,去创建一个
[1] Yoshifusa Ito,et al. Approximation of continuous functions on Rd by linear combinations of shifted rotations of a sigmoid function with and without scaling , 1992, Neural Networks.
[2] Allan Pinkus,et al. Multilayer Feedforward Networks with a Non-Polynomial Activation Function Can Approximate Any Function , 1991, Neural Networks.
[3] Ken-ichi Funahashi,et al. On the approximate realization of continuous mappings by neural networks , 1989, Neural Networks.
[4] H. A. Babri,et al. General approximation theorem on feedforward networks , 1997, Proceedings of ICICS, 1997 International Conference on Information, Communications and Signal Processing. Theme: Trends in Information Systems Engineering and Wireless Multimedia Communications (Cat..
[5] Chee Kheong Siew,et al. Extreme learning machine: Theory and applications , 2006, Neurocomputing.
[6] Kurt Hornik,et al. Multilayer feedforward networks are universal approximators , 1989, Neural Networks.
[7] Yoshifusa Ito,et al. Representation of functions by superpositions of a step or sigmoid function and their applications to neural network theory , 1991, Neural Networks.
[8] M. Viberg,et al. Adaptive neural nets filter using a recursive Levenberg-Marquardt search direction , 1998, Conference Record of Thirty-Second Asilomar Conference on Signals, Systems and Computers (Cat. No.98CH36284).
[9] Christopher J. Merz,et al. UCI Repository of Machine Learning Databases , 1996 .
[10] Tommy W. S. Chow,et al. Feedforward networks training speed enhancement by optimal initialization of the synaptic coefficients , 2001, IEEE Trans. Neural Networks.
[11] G. Lewicki,et al. Approximation by Superpositions of a Sigmoidal Function , 2003 .
[12] Bernard Widrow,et al. Improving the learning speed of 2-layer neural networks by choosing initial values of the adaptive weights , 1990, 1990 IJCNN International Joint Conference on Neural Networks.
[13] Anastasios N. Venetsanopoulos,et al. Artificial neural networks - learning algorithms, performance evaluation, and applications , 1992, The Kluwer international series in engineering and computer science.
[14] Guang-Bin Huang,et al. Upper bounds on the number of hidden neurons in feedforward networks with arbitrary bounded nonlinear activation functions , 1998, IEEE Trans. Neural Networks.
[15] Kurt Hornik,et al. Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.
[16] D. Serre. Matrices: Theory and Applications , 2002 .
[17] Hong Chen,et al. Approximation capability in C(R¯n) by multilayer feedforward networks and related problems , 1995, IEEE Trans. Neural Networks.
[18] Antonio J. Serrano,et al. A low-complexity fuzzy activation function for artificial neural networks , 2003, IEEE Trans. Neural Networks.
[19] Tahseen Ahmed Jilani,et al. A Comparison of First and Second Order Training Algorithms for Artificial Neural Networks , 2007, International Conference on Computational Intelligence.