Kolmogorov's spline network
暂无分享,去创建一个
[1] Yoh-Han Pao,et al. The ensemble approach to neural-network learning and generalization , 1999, IEEE Trans. Neural Networks.
[2] David A. Sprecher,et al. A Numerical Implementation of Kolmogorov's Superpositions , 1996, Neural Networks.
[3] Vladik Kreinovich,et al. Guaranteed Intervals for Kolmogorov’s Theorem (and Their Possible Relation to Neural Networks) , 2004 .
[4] Y. Makovoz. Random Approximants and Neural Networks , 1996 .
[5] David A. Sprecher,et al. A Numerical Implementation of Kolmogorov's Superpositions II , 1996, Neural Networks.
[6] Massood Tabib-Azar,et al. A net with complex weights , 2001, IEEE Trans. Neural Networks.
[7] Francesco Piazza,et al. Multilayer feedforward networks with adaptive spline activation function , 1999, IEEE Trans. Neural Networks.
[8] Francesco Piazza,et al. Learning and Approximation Capabilities of Adaptive Spline Activation Function Neural Networks , 1998, Neural Networks.
[9] Paolo Campolucci,et al. Complex-valued neural networks with adaptive spline activation function for digital-radio-links nonlinear equalization , 1999, IEEE Trans. Signal Process..
[10] Lionel Tarassenko,et al. Estimations of error bounds for neural-network function approximators , 1999, IEEE Trans. Neural Networks.
[11] D. Sprecher. On the structure of continuous functions of several variables , 1965 .
[12] M. Nees. Approximative versions of Kolmogorov's superposition theorem, proved constructively , 1994 .
[13] C. A. Hall,et al. On error bounds for spline interpolation , 1968 .
[14] Geoffrey E. Hinton,et al. Learning internal representations by error propagation , 1986 .
[15] J. L. Walsh,et al. The theory of splines and their applications , 1969 .
[16] R. Schapire. The Strength of Weak Learnability , 1990, Machine Learning.
[17] V. Tikhomirov. On the Representation of Continuous Functions of Several Variables as Superpositions of Continuous Functions of one Variable and Addition , 1991 .
[18] G. Lorentz,et al. Constructive approximation : advanced problems , 1996 .
[19] P. M. Prenter. Splines and variational methods , 1975 .
[20] A. Barron. Approximation and Estimation Bounds for Artificial Neural Networks , 1991, COLT '91.
[21] Vra Krkov. Kolmogorov's Theorem Is Relevant , 1991, Neural Computation.
[22] Vera Kurková,et al. Kolmogorov's theorem and multilayer neural networks , 1992, Neural Networks.
[23] Leo Breiman,et al. Hinging hyperplanes for regression, classification, and function approximation , 1993, IEEE Trans. Inf. Theory.
[24] Andrew R. Barron,et al. Universal approximation bounds for superpositions of a sigmoidal function , 1993, IEEE Trans. Inf. Theory.
[25] Ah Chung Tsoi,et al. Universal Approximation Using Feedforward Neural Networks: A Survey of Some Existing Methods, and Some New Results , 1998, Neural Networks.
[26] L. K. Jones,et al. Good weights and hyperbolic kernels for neural networks, projection pursuit, and pattern classification: Fourier strategies for extracting information from high-dimensional data , 1994, IEEE Trans. Inf. Theory.
[27] Andrew R. Barron,et al. Approximation and estimation bounds for artificial neural networks , 2004, Machine Learning.
[28] Chuanyi Ji,et al. Combinations of Weak Classifiers , 1996, NIPS.
[29] Tomaso A. Poggio,et al. Representation Properties of Networks: Kolmogorov's Theorem Is Irrelevant , 1989, Neural Computation.