Parsing Heuristic and Forward Search in First-Graders' Game-Play Behavior

Seventy-three children between 6 and 7 years of age were presented with a problem having ambiguous subgoal ordering. Performance in this task showed reliable fingerprints: (a) a non-monotonic dependence of performance as a function of the distance between the beginning and the end-states of the problem, (b) very high levels of performance when the first move was correct, and (c) states in which accuracy of the first move was significantly below chance. These features are consistent with a non-Markov planning agent, with an inherently inertial decision process, and that uses heuristics and partial problem knowledge to plan its actions. We applied a statistical framework to fit and test the quality of a proposed planning model (Monte Carlo Tree Search). Our framework allows us to parse out independent contributions to problem-solving based on the construction of the value function and on general mechanisms of the search process in the tree of solutions. We show that the latter are correlated with children's performance on an independent measure of planning, while the former is highly domain specific.

[1]  Allen Newell,et al.  Report on a general problem-solving program , 1959, IFIP Congress.

[2]  O. J. Dunn Multiple Comparisons among Means , 1961 .

[3]  H. Levene Robust tests for equality of variances , 1961 .

[4]  J. Piaget The Grasp of Consciousness: Action and Concept in the Young Child , 1976 .

[5]  S. Holm A Simple Sequentially Rejective Multiple Test Procedure , 1979 .

[6]  D. Klahr,et al.  Formal assessment of problem-solving and planning processes in preschool children , 1981, Cognitive Psychology.

[7]  Jesfis Peral,et al.  Heuristics -- intelligent search strategies for computer problem solving , 1984 .

[8]  D. Klahr Solving Problems with Ambiguous Subgoal Ordering: Preschoolers' Performance. , 1985 .

[9]  H. Arkes,et al.  The Psychology of Sunk Cost , 1985 .

[10]  Martin L. Puterman,et al.  Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .

[11]  Peter Norvig,et al.  Artificial Intelligence: A Modern Approach , 1995 .

[12]  Drew McDermott,et al.  A Heuristic Estimator for Means-Ends Analysis in Planning , 1996, AIPS.

[13]  S. Dehaene,et al.  The Number Sense: How the Mind Creates Mathematics. , 1998 .

[14]  B. Jones BOUNDED RATIONALITY , 1999 .

[15]  Bernhard Nebel,et al.  The FF Planning System: Fast Plan Generation Through Heuristic Search , 2011, J. Artif. Intell. Res..

[16]  D. Cox Karl Pearson and the Chi-Squared Test , 2002 .

[17]  Frédérick Garcia,et al.  On-Line Search for Solving Markov Decision Processes via Heuristic Sampling , 2004, ECAI.

[18]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[19]  Csaba Szepesvári,et al.  Bandit Based Monte-Carlo Planning , 2006, ECML.

[20]  M. Roesch,et al.  Dopamine neurons encode the better option in rats deciding between differently delayed or sized rewards , 2007, Nature Neuroscience.

[21]  Hector Geffner,et al.  Heuristics for Planning with Action Costs Revisited , 2008, ECAI.

[22]  Hector Geffner Heuristics, Planning and Cognition , 2010 .

[23]  T. McCormack,et al.  Planning in young children: A review and synthesis , 2011 .

[24]  Simon M. Lucas,et al.  A Survey of Monte Carlo Tree Search Methods , 2012, IEEE Transactions on Computational Intelligence and AI in Games.

[25]  Mariano Sigman,et al.  Mate Marote: A flexible automated framework for large-scale educational interventions , 2013, Comput. Educ..

[26]  M. Sigman,et al.  Training Planning and Working Memory in Third Graders , 2013 .

[27]  Diego E. Shalom,et al.  Far transfer to language and math of a short software-based gaming intervention , 2014, Proceedings of the National Academy of Sciences.