Learning capability and storage capacity of two-hidden-layer feedforward networks
暂无分享,去创建一个
[1] Vera Kurková,et al. Kolmogorov's theorem and multilayer neural networks , 1992, Neural Networks.
[2] Akito Sakurai. Tight Bounds for the VC-Dimension of Piecewise Polynomial Networks , 1998, NIPS.
[3] Masami Yamasaki,et al. The Lower Bound of the Capacity for a Neural Network with Multiple Hidden Layers , 1993 .
[4] Peter L. Bartlett,et al. Neural Network Learning - Theoretical Foundations , 1999 .
[5] Allan Pinkus,et al. Multilayer Feedforward Networks with a Non-Polynomial Activation Function Can Approximate Any Function , 1991, Neural Networks.
[6] Guang-Bin Huang,et al. Upper bounds on the number of hidden neurons in feedforward networks with arbitrary bounded nonlinear activation functions , 1998, IEEE Trans. Neural Networks.
[7] Tomaso A. Poggio,et al. Extensions of a Theory of Networks for Approximation and Learning , 1990, NIPS.
[8] Guang-Bin Huang,et al. Comments on "Approximation capability in C(Rn) by multilayer feedforward networks and related problems" , 1998, IEEE Trans. Neural Networks.
[9] Gary G. R. Green,et al. Neural networks, approximation theory, and finite precision computation , 1995, Neural Networks.
[10] Ken-ichi Funahashi,et al. On the approximate realization of continuous mappings by neural networks , 1989, Neural Networks.
[11] Marek Karpinski,et al. Polynomial bounds for VC dimension of sigmoidal neural networks , 1995, STOC '95.
[12] John Shawe-Taylor,et al. Sample sizes for multiple-output threshold networks , 1991 .
[13] John Shawe-Taylor. Sample sizes for sigmoidal neural networks , 1995, COLT '95.
[14] Eduardo D. Sontag,et al. Neural Networks with Quadratic VC Dimension , 1995, J. Comput. Syst. Sci..
[15] Eric B. Baum,et al. On the capabilities of multilayer perceptrons , 1988, J. Complex..
[16] Peter L. Bartlett,et al. Almost Linear VC-Dimension Bounds for Piecewise Polynomial Networks , 1998, Neural Computation.
[17] Yoshifusa Ito,et al. Approximation of continuous functions on Rd by linear combinations of shifted rotations of a sigmoid function with and without scaling , 1992, Neural Networks.
[18] A. Sakurai,et al. Tighter bounds of the VC-dimension of three layer networks , 1993 .
[19] Hong Chen,et al. Approximation capability in C(R¯n) by multilayer feedforward networks and related problems , 1995, IEEE Trans. Neural Networks.
[20] H. White,et al. Universal approximation using feedforward networks with non-sigmoid hidden layer activation functions , 1989, International 1989 Joint Conference on Neural Networks.
[21] Chong-Ho Choi,et al. Constructive neural networks with piecewise interpolation capabilities for function approximations , 1994, IEEE Trans. Neural Networks.
[22] Vladik Kreinovich,et al. Arbitrary nonlinearity is sufficient to represent all functions by neural networks: A theorem , 1991, Neural Networks.
[23] Peter L. Bartlett,et al. For Valid Generalization the Size of the Weights is More Important than the Size of the Network , 1996, NIPS.
[24] David Haussler,et al. What Size Net Gives Valid Generalization? , 1989, Neural Computation.
[25] Eduardo D. Sontag,et al. UNIQUENESS OF WEIGHTS FOR NEURAL NETWORKS , 1993 .
[26] H. White,et al. There exists a neural network that does not make avoidable mistakes , 1988, IEEE 1988 International Conference on Neural Networks.
[27] Peter L. Bartlett,et al. The Sample Complexity of Pattern Classification with Neural Networks: The Size of the Weights is More Important than the Size of the Network , 1998, IEEE Trans. Inf. Theory.
[28] Shin'ichi Tamura,et al. Capabilities of a four-layered feedforward neural network: four layers versus three , 1997, IEEE Trans. Neural Networks.
[29] Kurt Hornik,et al. Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.