Inter-Rater Reliability of the Strain Index

The Strain Index is one of several tools available to evaluate exposure to musculoskeletal stressors in the workplace in order to predict whether workers are at an increased risk of developing distal upper extremity disorders. The purpose of this study was to determine the inter-rater reliability of the Strain Index. Fifteen raters initially estimated or measured data for five Strain Index task variables: (1) intensity of exertion, (2) duration of exertion, (3) efforts per minute, (4) hand/wrist posture, and (5) speed of work, using a computer to view 61 video segments of single task jobs. Video segments were organized to provide a balanced number for each of the five variables under study, where each segment was used to provide data for a single variable. Raters then assigned rating values to each task according to published procedures. For an additional 12 segments, raters performed complete Strain Index analyses, including calculation of Strain Index scores, which were used to determine hazard classifications. Raters assessed the tasks both as individuals and as members of five three-person teams. Inter-rater reliabilities of the procedures leading to and including the Strain Index scores were assessed using an intraclass correlation coefficient, (ICC(2,1)). Inter-rater reliability for the dichotomous hazard classification was assessed using Kuder-Richardson-20 (KR-20, an ICC for dichotomous data). For task variables and estimated data, ICC(2,1) varied between 0.66–0.84 for individuals and 0.48–0.93 for teams. The Strain Index score had an ICC(2,1) of 0.43 and 0.64 for individuals and teams, respectively. For the most important variable, hazard classification, KR-20 was 0.91 for the individuals and 0.89 for the teams.

[1]  D. Streiner,et al.  Health Measurement Scales: A practical guide to thier development and use , 1989 .

[2]  Douglas P Gross,et al.  Reliability of safe maximum lifting determinations of a functional capacity evaluation. , 2002, Physical therapy.

[3]  J. Uebersax Validity inferences from interobserver agreement. , 1988 .

[4]  R. L. Ebel,et al.  Estimation of the reliability of ratings , 1951 .

[5]  L. Cronbach Coefficient alpha and the internal structure of tests , 1951 .

[6]  H. Brenner,et al.  Dependence of Weighted Kappa Coefficients on the Number of Categories , 1996, Epidemiology.

[7]  J. Fleiss The design and analysis of clinical experiments , 1987 .

[8]  J. R. Landis,et al.  The measurement of observer agreement for categorical data. , 1977, Biometrics.

[9]  Jacob Cohen A Coefficient of Agreement for Nominal Scales , 1960 .

[10]  M. W. Richardson,et al.  The theory of the estimation of test reliability , 1937 .

[11]  K. McGraw,et al.  Forming inferences about some intraclass correlation coefficients. , 1996 .

[12]  A Garg,et al.  Participatory ergonomics in a red meat packing plant. Part II: Case studies. , 1997, American Industrial Hygiene Association journal.

[13]  Ronald A. Berk,et al.  An Analysis of Variance Model for Assessing Reliability of Naturalistic Observations , 1978 .

[14]  G. Borg Psychophysical bases of perceived exertion. , 1982, Medicine and science in sports and exercise.

[15]  M. Szklo,et al.  Epidemiology: Beyond the Basics , 1999 .

[16]  Wendi Ann Latko Development and evaluation of an observational method for quantifying exposure to hand activity and other physical stressors in manual work. , 1997 .

[17]  R. F. Fagot,et al.  A generalized family of coefficients of relational agreement for numerical scales , 1993 .

[18]  John S. Uebersax,et al.  Diversity of decision-making models and the measurement of interrater agreement. , 1987 .

[19]  G. Maislin,et al.  Intra-observer and inter-observer variability in scoring laparoscopic diagnosis of pelvic adhesions. , 1995, Human reproduction.

[20]  L McAtamney,et al.  RULA: a survey method for the investigation of work-related upper limb disorders. , 1993, Applied ergonomics.

[21]  A. Garg,et al.  The Strain Index: a proposed method to analyze jobs for risk of distal upper extremity disorders. , 1995, American Industrial Hygiene Association journal.

[22]  W. Willett,et al.  Misinterpretation and misuse of the kappa statistic. , 1987, American journal of epidemiology.

[23]  Jacob Cohen,et al.  Weighted kappa: Nominal scale agreement provision for scaled disagreement or partial credit. , 1968 .

[24]  R A Berk,et al.  Generalizability of behavioral observations: a clarification of interobserver agreement and interobserver reliability. , 1979, American journal of mental deficiency.

[25]  J. Moore,et al.  Predictive Validity of the Strain Index in Turkey Processing , 2001, Journal of occupational and environmental medicine.

[26]  Jacob Cohen,et al.  The Equivalence of Weighted Kappa and the Intraclass Correlation Coefficient as Measures of Reliability , 1973 .

[27]  T J Armstrong,et al.  Reliability of physical examination of the upper extremity among keyboard operators. , 2000, American journal of industrial medicine.

[28]  Sally Jackson,et al.  Random Factors in ANOVA , 1993 .

[29]  J. Moore,et al.  Validity of generic risk factors and the strain index for predicting nontraumatic distal upper extremity morbidity. , 2001, AIHAJ : a journal for the science of occupational and environmental health and safety.

[30]  L Punnett,et al.  Evaluation of interrater reliability for posture observations in a field study. , 1999, Applied ergonomics.

[31]  J. Moore,et al.  Predictive validity of the strain index in manufacturing facilities. , 2002, Applied occupational and environmental hygiene.

[32]  P D Sampson,et al.  Measuring interrater reliability among multiple raters: an example of methods for nominal data. , 1990, Statistics in medicine.

[33]  S Hignett,et al.  Rapid entire body assessment (REBA). , 2000, Applied ergonomics.

[34]  J R Morrow,et al.  How "significant" is your reliability? , 1993, Research quarterly for exercise and sport.

[35]  J. Fleiss,et al.  Intraclass correlations: uses in assessing rater reliability. , 1979, Psychological bulletin.