Analysis of the rate of convergence of least squares neural network regression estimates in case of measurement errors

Estimation of a regression function from data which consists of an independent and identically distributed sample of the underlying distribution with additional measurement errors in the independent variables is considered. It is allowed that the measurement errors are not independent and have a nonzero mean. It is shown that the rate of convergence of suitably defined least squares neural network estimates applied to this data is similar to the rate of convergence of least squares neural network estimates applied to an independent and identically distributed sample of the underlying distribution as long as the measurement errors are small.

[1]  Adil M. Bagirov,et al.  An $L_{2}$ -Boosting Algorithm for Estimation of a Regression Function , 2010, IEEE Transactions on Information Theory.

[2]  George Cybenko,et al.  Approximation by superpositions of a sigmoidal function , 1992, Math. Control. Signals Syst..

[3]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[4]  A. Barron,et al.  Statistical properties of artificial neural networks , 1989, Proceedings of the 28th IEEE Conference on Decision and Control,.

[5]  Adam Krzyzak,et al.  A Distribution-Free Theory of Nonparametric Regression , 2002, Springer series in statistics.

[6]  Peter L. Bartlett,et al.  Neural Network Learning - Theoretical Foundations , 1999 .

[7]  Jianqing Fan,et al.  Nonparametric regression with errors in variables , 1993 .

[8]  L. Györfi,et al.  A Distribution-Free Theory of Nonparametric Regression (Springer Series in Statistics) , 2002 .

[9]  Gábor Lugosi,et al.  Nonparametric estimation via empirical risk minimization , 1995, IEEE Trans. Inf. Theory.

[10]  A. Krzyżak,et al.  Adaptive regression estimation with multilayer feedforward neural networks , 2005 .

[11]  D. Ruppert,et al.  Nonparametric regression in the presence of measurement error , 1999 .

[12]  Daniel F. McCaffrey,et al.  Convergence rates for single hidden layer feedforward networks , 1994, Neural Networks.

[13]  Andrew R. Barron,et al.  Complexity Regularization with Application to Artificial Neural Networks , 1991 .

[14]  Raymond J Carroll,et al.  A Design-Adaptive Local Polynomial Estimator for the Errors-in-Variables Problem , 2009, Journal of the American Statistical Association.

[15]  Jan Mielniczuk,et al.  Consistency of multilayer perceptron regression estimators , 1993, Neural Networks.

[16]  László Györfi,et al.  A Probabilistic Theory of Pattern Recognition , 1996, Stochastic Modelling and Applied Probability.

[17]  Ken-ichi Funahashi,et al.  On the approximate realization of continuous mappings by neural networks , 1989, Neural Networks.

[18]  George Cybenko,et al.  Approximation by superpositions of a sigmoidal function , 1989, Math. Control. Signals Syst..

[19]  Andrew R. Barron,et al.  Universal approximation bounds for superpositions of a sigmoidal function , 1993, IEEE Trans. Inf. Theory.

[20]  Yoshua Bengio,et al.  Pattern Recognition and Neural Networks , 1995 .

[21]  Alexander Meister,et al.  Nonparametric Regression Estimation in the Heteroscedastic Errors-in-Variables Problem , 2007 .

[22]  Heinz Bauer,et al.  Probability Theory , 2021, Foundations of Constructive Probability Theory.

[23]  Anders Krogh,et al.  Introduction to the theory of neural computation , 1994, The advanced book program.

[24]  Kurt Hornik,et al.  Some new results on neural network approximation , 1993, Neural Networks.

[25]  Shlomo Nir,et al.  NATO ASI Series , 1995 .

[26]  M. Kohler,et al.  Nonasymptotic Bounds on the L2 Error of Neural Network Regression Estimates , 2006 .