What to Make Of and How to Interpret Process Data

Maddox (2017) argues that respondents’ talk and gesture during an assessment inform researchers how a response product has evolved. Indeed, how a task is performed represents key information for psychological and educational assessment. In an ancient example: Gideon was required by the Lord to select those men who lap the water with their tongues, but not those who kneel down to drink (Judges 7:5 New International Version). In cognitive ability testing, process data can be defined as empirical information about the cognitive (as well as meta-cognitive, motivational, and affective) states and related behavior that mediate the effect of the measured construct(s) on the task product (i.e., item score). Thus, operationally, process data can be regarded as the empirical data reflecting the course of working on a test item. Recently, process data has increasingly gained attention in cognitive ability testing given the digitalization of measurement and the possibility of exploiting log file data. Other sources of process data are, for instance, concurrent think aloud protocols, screen capturing, eye tracking, facial expression, video-recorded behavior, and physiological sensor data (Azevedo, 2015). As shown by Maddox for large-scale assessments, even talk and gesture can be regarded as useful process data. In this case, the process data is not only video-recorded but also observed by the interviewer in situ; the interviewer interactively uses it to influence the test-taking process and to reduce construct-irrelevant variance. Thus, like product data (e.g., scores), process data is used to draw inferences. We argue in the following that the interpretation and use of process data and derived indicators require validation, just as product data do (Kane, 2013). This theoretical background, including some examples about log file data, sets the ground for our comments on Maddox’s use of “talk and gesture as process data.”

[1]  R. H. Klein Entink,et al.  A Multivariate Multilevel Approach to the Modeling of Accuracy and Speed of Test Takers , 2008, Psychometrika.

[2]  B. Maddox Talk and Gesture as Process Data , 2017 .

[3]  P. Ekman,et al.  What the face reveals : basic and applied studies of spontaneous expression using the facial action coding system (FACS) , 2005 .

[4]  Qiang Ji,et al.  An automated face reader for fatigue detection , 2004, Sixth IEEE International Conference on Automatic Face and Gesture Recognition, 2004. Proceedings..

[5]  Wim J. van der Linden,et al.  Bayesian Procedures for Identifying Aberrant Response-Time Patterns in Adaptive Testing , 2008 .

[6]  M. Kane Validating the Interpretations and Uses of Test Scores , 2013 .

[7]  David R. Traum,et al.  Embodied agents for multi-party dialogue in immersive virtual worlds , 2002, AAMAS '02.

[8]  James Baker,et al.  A historical perspective of speech recognition , 2014, CACM.

[9]  Arthur C. Graesser,et al.  AutoTutor and affective autotutor: Learning by talking with cognitively and emotionally intelligent computers that talk back , 2012, TIIS.

[10]  Yue Jia,et al.  Collecting, Analyzing, and Interpreting Response Time, Eye-Tracking, and Log Data , 2017 .

[11]  Steven L. Wise,et al.  Setting the Response Time Threshold Parameter to Differentiate Solution Behavior From Rapid-Guessing Behavior , 2007 .

[12]  L. Suchman,et al.  Interactional Troubles in Face-to-Face Survey Interviews , 1990 .

[13]  Oliver Lüdtke,et al.  Conditioning factors of test-taking engagement in PIAAC: an exploratory IRT modelling approach considering person and item characteristics , 2017 .

[14]  Richard A. Harshman,et al.  Indexing by Latent Semantic Analysis , 1990, J. Am. Soc. Inf. Sci..

[15]  R. Almond,et al.  A BRIEF INTRODUCTION TO EVIDENCE-CENTERED DESIGN , 2003 .

[16]  Daniela Ackermann-Piek,et al.  Interviewer Behavior and Interviewer Characteristics in PIAAC Germany , 2014 .

[17]  Yi-Hsuan Lee,et al.  Using response time to investigate students' test-taking behaviors in a NAEP computer-based study , 2014, Large-scale Assessments in Education.

[18]  Kristy Elizabeth Boyer,et al.  Automatically Recognizing Facial Expression: Predicting Engagement and Frustration , 2013, EDM.

[19]  D. Molenaar The Value of Response Times in Item Response Modeling , 2015 .

[20]  R. Azevedo Defining and Measuring Engagement and Learning in Science: Conceptual, Theoretical, Methodological, and Analytical Issues , 2015 .

[21]  D. Borsboom,et al.  The concept of validity. , 2004, Psychological review.

[22]  Samuel Greiff,et al.  Understanding students' performance in a computer-based assessment of complex problem solving: An analysis of behavioral data from computer-generated log files , 2016, Comput. Hum. Behav..

[23]  Validating Score Interpretations Based on Response Processes , 2017 .