GEMINI: Gradient Estimation Through Matrix Inversion After Noise Injection
暂无分享,去创建一个
Learning procedures that measure how random perturbations of unit activities correlate with changes in reinforcement are inefficient but simple to implement in hardware. Procedures like back-propagation (Rumelhart, Hinton and Williams, 1986) which compute how changes in activities affect the output error are much more efficient, but require more complex hardware. GEMINI is a hybrid procedure for multilayer networks, which shares many of the implementation advantages of correlational reinforcement procedures but is more efficient. GEMINI injects noise only at the first hidden layer and measures the resultant effect on the output error. A linear network associated with each hidden layer iteratively inverts the matrix which relates the noise to the error change, thereby obtaining the error-derivatives. No back-propagation is involved, thus allowing unknown non-linearities in the system. Two simulations demonstrate the effectiveness of GEMINI.
[1] P. Anandan,et al. Pattern-recognizing stochastic learning automata , 1985, IEEE Transactions on Systems, Man, and Cybernetics.
[2] D. Rumelhart. Learning internal representations by back-propagating errors , 1986 .
[3] Yann LeCun,et al. Modeles connexionnistes de l'apprentissage , 1987 .
[4] Yann LeCun,et al. Improving the convergence of back-propagation learning with second-order methods , 1989 .