Mind Reading by Machine Learning: A Doubly Bayesian Method for Inferring Mental Representations

Mind Reading by Machine Learning: A Doubly Bayesian Method for Inferring Mental Representations Ferenc Husz´ ar (fh277@eng.cam.ac.uk ) Computational and Biological Learning Lab, Dept Engineering, U Cambridge, Cambridge CB2 1PZ, UK Uta Noppeney (uta.noppeney@tuebingen.mpi.de) Max Planck Institute for Biological Cybernetics, Spemannstrasse 41, T¨ ubingen 72076, Germany M´ at´ e Lengyel (m.lengyel@eng.cam.ac.uk) Computational and Biological Learning Lab, Dept Engineering, U Cambridge, Cambridge CB2 1PZ, UK Abstract A central challenge in cognitive science is to measure and quantify the mental representations humans develop – in other words, to ‘read’ subject’s minds. In order to elimi- nate potential biases in reporting mental contents due to verbal elaboration, subjects’ responses in experiments are often limited to binary decisions or discrete choices that do not require conscious reflection upon their mental contents. However, it is unclear what such impoverished data can tell us about the potential richness and dy- namics of subjects’ mental representations. To address this problem, we used ideal observer models that for- malise choice behaviour as (quasi-)Bayes-optimal, given subjects’ representations in long-term memory, acquired through prior learning, and the stimuli currently avail- able to them. Bayesian inversion of such ideal observer models allowed us to infer subjects’ mental representation from their choice behaviour in a variety of psychophysical tasks. The inferred mental representations also allowed us to predict future choices of subjects with reasonable accuracy, even in tasks that were different from those in which the representations were estimated. These results demonstrate a significant potential in standard binary decision tasks to recover detailed information about sub- jects’ mental representations. Introduction Cognitive science studies the mental representations hu- mans (and other animals) develop and the way these representations are used to perform particular tasks. A central challenge is to measure and quantify such men- tal representations experimentally – in other words, to ‘read’ subjects’ minds. A classical approach to this is to ask subjects directly to report their mental contents verbally. Unfortunately, this procedure is prone to intro- ducing biases arising from verbal processing, and from the educational and cultural backgrounds of subjects (Ericsson & Simon, 1980; Russo et al., 1989). In order to eliminate these biases, an alternative approach is to limit subjects’ responses to simple binary decisions or discrete choices that do not require conscious reflection upon their mental contents. However, it is unclear what such impoverished data can tell us about the potential richness and dynamics of subjects’ mental contents. A powerful computational framework formalises the goal of learning as estimating the probability distribution or density of stimuli (Hinton & Sejnowski, 1986; Dayan & Abbott, 2001). This motivates many formal theories of human learning and cognition to model the relevant mental content of a subject either implicitly or explicitly as a ‘subjective’ distribution over possible stimuli (Chater et al., 2006; Sanborn & Griffiths, 2008). In this study we adopted this representation, and our goal was to estimate subjects’ subjective distributions solely from their responses in simple binary decision tasks without making any assumptions about the process by which those subjective distributions were acquired, i. e. learning. Ideal observer models are widely used for explaining human behaviour in various psychophysics tasks (Geisler, 2003). They formalise (quasi-)optimal decision making strategies given the information available to subjects and their background knowledge about the task, which in our case includes their subjective distributions. While previous studies mostly used ideal observer models to de- termine optimal performance in particular tasks to which human performance could then be compared, we treat them as stochastic models formalising the link between subjective distributions (the unobserved variable), and test stimuli and responses (the observed variables). Our main observation is that such models can be used to provide the likelihood in a Bayesian statistical analysis of subjective distributions, thus enabling one to infer mental contents from task responses in a principled way. We term our approach doubly Bayesian, as we assume that subjects act as quasi-ideal observers, which entails Bayesian inference on their side; and then we use these ideal-observer models in a Bayesian framework to infer a posterior distribution of possible subjective distributions. Inferring subjective distributions The graphical model (Koller & Friedman, 2009) in Fig. 1A describes our model of a subject’s behaviour in a session of a psychophysics experiment. We assume that the sub- ject entertains a subjective distribution P over possible stimuli, and that this distribution does not change over the analysed session. In trial i of the experiment, the subject is presented a set of test stimuli S i and gives a re- sponse r i . The value of r i depends on the current stimuli S i , the subjective distribution P, and ‘link’ parameters Θ O describing further aspects of observation and decision making, such as attention, perceptual noise, etc.

[1]  K. A. Ericsson,et al.  Verbal reports as data. , 1980 .

[2]  Geoffrey E. Hinton,et al.  Learning and relearning in Boltzmann machines , 1986 .

[3]  Eric J. Johnson,et al.  The validity of verbal protocols , 1989, Memory & cognition.

[4]  C. Gallistel The organization of learning , 1990 .

[5]  Nir Vulkan An Economist's Perspective on Probability Matching , 2000 .

[6]  W. Geisler Ideal Observer Analysis , 2002 .

[7]  S. Kakade,et al.  Acquisition and extinction in autoshaping. , 2002, Psychological review.

[8]  C. Gallistel,et al.  The learning curve: implications of a quantitative analysis. , 2004, Proceedings of the National Academy of Sciences of the United States of America.

[9]  Joshua B. Tenenbaum,et al.  A Generative Theory of Similarity , 2005 .

[10]  Liam Paninski,et al.  Nonparametric inference of prior probabilities from Bayes-optimal behavior , 2005, NIPS.

[11]  Emery N Brown,et al.  Analysis and design of behavioral experiments to characterize population learning. , 2005, Journal of neurophysiology.

[12]  Philipp Slusallek,et al.  Introduction to real-time ray tracing , 2005, SIGGRAPH Courses.

[13]  J. Tenenbaum,et al.  Special issue on “Probabilistic models of cognition , 2022 .

[14]  D. Sagi,et al.  Inverse modeling of human contrast response , 2007, Vision Research.

[15]  Thomas L. Griffiths,et al.  Markov Chain Monte Carlo with People , 2007, NIPS.

[16]  Richard N Aslin,et al.  Bayesian learning of visual chunks by human observers , 2008, Proceedings of the National Academy of Sciences.

[17]  Carl E. Rasmussen,et al.  Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.

[18]  Nir Friedman,et al.  Probabilistic Graphical Models - Principles and Techniques , 2009 .

[19]  Misha Tsodyks,et al.  Mapping dynamic memories of gradually changing objects , 2009, Proceedings of the National Academy of Sciences.

[20]  Andrew Gelman,et al.  Handbook of Markov Chain Monte Carlo , 2011 .

[21]  Radford M. Neal MCMC Using Hamiltonian Dynamics , 2011, 1206.1901.