EasyDIAg: A tool for easy determination of interrater agreement

Reliable measurements are fundamental for the empirical sciences. In observational research, measurements often consist of observers categorizing behavior into nominal-scaled units. Since the categorization is the outcome of a complex judgment process, it is important to evaluate the extent to which these judgments are reproducible, by having multiple observers independently rate the same behavior. A challenge in determining interrater agreement for timed-event sequential data is to develop clear objective criteria to determine whether two raters’ judgments relate to the same event (the linking problem). Furthermore, many studies presently report only raw agreement indices, without considering the degree to which agreement can occur by chance alone. Here, we present a novel, free, and open-source toolbox (EasyDIAg) designed to assist researchers with the linking problem, while also providing chance-corrected estimates of interrater agreement. Additional tools are included to facilitate the development of coding schemes and rater training.

[1]  W. Deming,et al.  On a Least Squares Adjustment of a Sampled Frequency Table When the Expected Marginal Totals are Known , 1940 .

[2]  Patsy Haccou,et al.  Statistical Analysis of Behavioural Data: An Approach Based on Time-structured Models , 1992 .

[3]  Lucas P J J Noldus,et al.  Reliability analysis of observational data: Problems, solutions, and software implementation , 2003, Behavior research methods, instruments, & computers : a journal of the Psychonomic Society, Inc.

[4]  Roger Bakeman,et al.  SDIS: A sequential data interchange standard , 1992 .

[5]  S. Reeves,et al.  Discourse Analysis , 2018, Understanding Communication Research Methods.

[6]  T. Taris,et al.  Measuring the Agreement between Sequences , 1995 .

[7]  Robert Rosenthal,et al.  Can you believe my eyes? The importance of interobserver reliability statistics in observations of animal behaviour , 2009, Animal Behaviour.

[8]  Janet Beavin Bavelas,et al.  Gesturing on the telephone: Independent effects of dialogue and visibility. , 2008 .

[9]  A. Feinstein,et al.  High agreement but low kappa: II. Resolving the paradoxes. , 1990, Journal of clinical epidemiology.

[10]  Roger Bakeman,et al.  Understanding Log-linear Analysis With Ilog: An Interactive Approach , 1994 .

[11]  M. Potter,et al.  Resolving the paradoxes , 2008 .

[12]  Peter J. Rousseeuw,et al.  Finding Groups in Data: An Introduction to Cluster Analysis , 1990 .

[13]  Roger Bakeman,et al.  Observer agreement for timed-event sequential data: A comparison of time-based and event-based algorithms , 2009, Behavior research methods.

[14]  Andrew Mackinnon,et al.  A spreadsheet for the calculation of comprehensive statistics for the assessment of diagnostic tests and inter-rater agreement , 2000, Comput. Biol. Medicine.

[15]  Hedda Lausberg,et al.  Understanding Body Movement , 2014 .

[16]  Jacob Cohen A Coefficient of Agreement for Nominal Scales , 1960 .

[17]  Hedda Lausberg,et al.  Understanding Body Movement : A Guide to Empirical Research on Nonverbal Behaviour - With an Introduction to the NEUROGES Coding System , 2013 .

[18]  R. Bakeman,et al.  Sequential analysis and observational methods for the behavioral sciences , 2011 .

[19]  Roger Bakeman,et al.  Observer agreement for event sequences: Methods and software for sequence alignment and reliability estimates , 2007, Behavior research methods.

[20]  Hedda Lausberg,et al.  Methods in Gesture Research: , 2009 .

[21]  D. McNeill Hand and Mind: What Gestures Reveal about Thought , 1992 .

[22]  G. Miller,et al.  Handbook of Interpersonal Communication , 1985 .