Comparing Biases for Minimal Network Construction with Back-Propagation
暂无分享,去创建一个
Rumelhart (1987), has proposed a method for choosing minimal or "simple" representations during learning in Back-propagation networks. This approach can be used to (a) dynamically select the number of hidden units, (b) construct a representation that is appropriate for the problem and (c) thus improve the generalization ability of Back-propagation networks. The method Rumelhart suggests involves adding penalty terms to the usual error function. In this paper we introduce Rumelhart's minimal networks idea and compare two possible biases on the weight search space. These biases are compared in both simple counting problems and a speech recognition problem. In general, the constrained search does seem to minimize the number of hidden units required with an expected increase in local minima.
[1] N. Draper,et al. Applied Regression Analysis , 1966 .
[2] C. D. Gelatt,et al. Optimization by Simulated Annealing , 1983, Science.
[3] Geoffrey E. Hinton,et al. Learning internal representations by error propagation , 1986 .