Learning automata processing ergodicity of the mean: The two-action case

Learning automata which update their action probabilities on the basis of the responses they get from an environment are considered. The automata update the probabilities whether the environment responds with a reward or a penalty. An automation is said to possess ergodicity of the mean (EM) if the mean action probability is the total state probability of an ergodic Markov chain. The only known EM algorithm is the linear reward-penalty (LRP) scheme. For the two-action case, necessary and sufficient conditions have been derived for nonlinear updating schemes to be EM. The method of controlling the rate of convergence of this scheme is presented. In particular, a generalized linear algorithm has been proposed which is superior to the LRP scheme. The expression for the variance of the limiting action probabilities of this scheme is derived.