Learning with Noise and Regularizers in Multilayer Neural Networks
暂无分享,去创建一个
We study the effect of noise and regularization in an on-line gradient-descent learning scenario for a general two-layer student network with an arbitrary number of hidden units. Training examples are randomly drawn input vectors labeled by a two-layer teacher network with an arbitrary number of hidden units; the examples are corrupted by Gaussian noise affecting either the output or the model itself. We examine the effect of both types of noise and that of weight-decay regularization on the dynamical evolution of the order parameters and the generalization error in various phases of the learning process.
[1] Klaus Schulten,et al. A Numerical Study on Learning Curves in Stochastic Multilayer Feedforward Networks , 1996, Neural Computation.
[2] M. V. Rossum,et al. In Neural Computation , 2022 .
[3] Heekuck Oh,et al. Neural Networks for Pattern Recognition , 1993, Adv. Comput..
[4] B. M. Fulk. MATH , 1992 .