Action-Reinforcement Learning Versus Rule Learning

How well do various learning models predict the dynamics of the population distribution of play in a variety of games? We measure and compare the in-sample and out-of-sample prediction performance of seven action-reinforcement learning models as well as Rule Learning for symmetric normal-form games. The criteria include log-likelihood values and root-mean-squared error. We conclude that a logit best-reply dynamic with inertia and adaptive expectations (LBRIAE) is the best among the action-reinforcement models. We formulate a population Rule Learning model that nests the LBRIAE model and find that Rule Learning is statistically superior. Sophistication is learned in the sense that random behavior and herd behavior decline substantially, and level-1 and level-2 behaviors increase over time.

[1]  Miguel A. Costa-Gomes,et al.  Cognition and Behavior in Normal-Form Games: An Experimental Study , 1998 .

[2]  R. Duncan Luce,et al.  Individual Choice Behavior , 1959 .

[3]  Dale O. Stahl,et al.  Population Rule Learning in Symmetric Normal-Form Games , 2001 .

[4]  R. Rob,et al.  Learning, Mutation, and Long Run Equilibria in Games , 1993 .

[5]  John Duffy,et al.  On the Robustness of Behaviour in Experimental ‘Beauty Contest’ Games , 1997 .

[6]  Dilip Mookherjee,et al.  Learning and Decision Costs in Experimental Constant Sum Games , 1997 .

[7]  A. Rapoport,et al.  Organizational Behavior and Human Decision Processes Randomization and Adaptive Learning in a Simplified Poker Game , 2022 .

[8]  Dilip Mookherjee,et al.  Learning behavior in an experimental matching pennies game , 1994 .

[9]  Ying-Wong Cheung,et al.  Individual Learning in Games: Some Laboratory Results , 1995 .

[10]  Nick Feltovich,et al.  Reinforcement-based vs. Belief-based Learning Models in Experimental Asymmetric-information Games , 2000 .

[11]  A. Roth,et al.  Learning in Extensive-Form Games: Experimental Data and Simple Dynamic Models in the Intermediate Term* , 1995 .

[12]  R. Selten Evolution, learning, and economic behavior , 1991 .

[13]  Colin Camerer,et al.  Sophisticated EWA Learning and Strategic Teaching in Repeated Games , 2000 .

[14]  R. Lewontin ‘The Selfish Gene’ , 1977, Nature.

[15]  D. Stahl,et al.  On Players' Models of Other Players: Theory and Experimental Evidence , 1995 .

[16]  Dale O. Stahl,et al.  Rule Learning in Symmetric Normal-Form Games: Theory and Evidence , 2000, Games Econ. Behav..

[17]  D. Fudenberg,et al.  The Theory of Learning in Games , 1998 .

[18]  J. Huyck,et al.  Strategic Uncertainty, Equilibrium Selection, and Coordination Failure in Average Opinion Games , 1991 .

[19]  Colin Camerer,et al.  Experience‐weighted Attraction Learning in Normal Form Games , 1999 .

[20]  Michael M. Cohen,et al.  A comparison of learning models , 1995 .

[21]  Fang-Fang Tang,et al.  Anticipatory learning in two-person games: some experimental results , 2001 .

[22]  Josef Hofbauer,et al.  The theory of evolution and dynamical systems , 1988 .

[23]  Nagel,et al.  Experimental Results on the Centipede Game in Normal Form: An Investigation on Learning. , 1998, Journal of mathematical psychology.

[24]  V. Crawford Adaptive dynamics in coordination games , 1995 .

[25]  F. Tang,et al.  Learning and incentive‐Compatible Mechanisms for Public Goods Provision: An Experimental Study , 1998, Journal of Political Economy.

[26]  Daniel Friedman,et al.  A comparison of learning and replicator dynamics using experimental data , 1998 .

[27]  R. Battalio,et al.  Selection Dynamics, Asymptotic Stability, and Adaptive Behavior , 1994, Journal of Political Economy.

[28]  A. Roth,et al.  Predicting How People Play Games: Reinforcement Learning in Experimental Games with Unique, Mixed Strategy Equilibria , 1998 .

[29]  Dale O. Stahl,et al.  Aspiration-based and reciprocity-based rules in learning dynamics for symmetric normal-form games , 2002 .

[30]  D. Stahl Boundedly rational rule learning in a guessing game , 1996 .

[31]  Milton Friedman,et al.  Essays in Positive Economics , 1954 .