Observer agreement on pen level prevalence of clinical signs in finishing pigs.

Four observers performed a standard clinical examination of finisher pigs in two commercial finisher herds. In herd 1,600 finisher pigs in 44 pens were examined. The observers assessed clinical signs of lameness, umbilical hernia and tail bite according to a standardized procedure. The prevalence of the clinical signs was estimated at the pen level. The procedure was repeated after 3 months in another herd, where 730 finisher pigs in 69 pens were examined. The agreement between observer pairs was assessed using prevalence-adjusted bias-adjusted kappa (PABAK) and proportionate-agreement estimates (Ppo) and Pneg). Observer bias was present despite training and standardization of the participating observers. The highest pen level agreement for the observer pairs was found for pens that had one or more pigs with tail bite (PABAK = 0.82-1.00) and umbilical hernia (PABAK = 0.77-1.00). The agreement was fair-to-moderate for identification of pens holding one or more lame pigs (PABAK = 0.27-0.7). In general, the average agreement of observer pairs on absence of clinical signs (Pneg) was higher than for presence (Ppos). The observer bias varied between observer pairs and with the clinical signs.

[1]  A. Feinstein,et al.  High agreement but low kappa: I. The problems of two paradoxes. , 1990, Journal of clinical epidemiology.

[2]  H Stryhn,et al.  Conditional dependence between tests affects the diagnosis and surveillance of animal diseases. , 2000, Preventive veterinary medicine.

[3]  M. Costantini,et al.  Interobserver agreement in the histologic diagnosis of colorectal polyps. the experience of the multicenter adenoma colorectal study (SMAC). , 2003, Journal of clinical epidemiology.

[4]  T. Fujinaga,et al.  Equine haptoglobin: isolation, characterization, and the effects of ageing, delivery and inflammation on its serum concentration. , 1992, The Journal of veterinary medical science.

[5]  I J Lean,et al.  Within and between observer agreement on ultrasonic evaluation of bovine ovarian structures. , 1992, Australian veterinary journal.

[6]  Henson De Studies on observer variation. Should the rules be changed , 1991 .

[7]  A. Feinstein,et al.  Spectrum Bias in the Evaluation of Diagnostic Tests: Lessons from the Rapid Dipstick Test for Urinary Tract Infection , 1992, Annals of Internal Medicine.

[8]  J. Lowe,et al.  Observer variation in equine abdominal auscultation. , 1990, Equine veterinary journal.

[9]  J. R. Landis,et al.  The measurement of observer agreement for categorical data. , 1977, Biometrics.

[10]  L. Roberts,et al.  Bovine haptoglobin response in clinically defined field conditions , 1991, Veterinary Record.

[11]  P. Tugwell,et al.  An experiment in reducing interobserver variability of the examination for joint tenderness. , 1988, The Journal of rheumatology.

[12]  P Sarmandal,et al.  A comparison of three methods of assessing interobserver variation applied to measurement of the symphysis‐fundal height , 1989, British journal of obstetrics and gynaecology.

[13]  G. Theron,et al.  Inter-observer variation in symphysis-fundus measurements. A plea for individualised antenatal care. , 1989, South African medical journal = Suid-Afrikaanse tydskrif vir geneeskunde.

[14]  R. Morton The less acceptable face of bias , 2000, The Lancet.

[15]  B. Everitt,et al.  Large sample standard errors of kappa and weighted kappa. , 1969 .

[16]  L. Green,et al.  Study of clinical mastitis in British dairy herds with bulk milk somatic cell counts less than 150,000 cells/ml , 2002, Veterinary Record.

[17]  J. Carlin,et al.  Bias, prevalence and kappa. , 1993, Journal of clinical epidemiology.

[18]  J. Kent,et al.  Assessment of an immunoturbidimetric method for measuring equine serum haptoglobin concentrations. , 1991, Equine veterinary journal.

[19]  H. Kraemer,et al.  Measurement of reliability for categorical data in medical research , 1992, Statistical methods in medical research.

[20]  A. Silman,et al.  Statistical methods for assessing observer variability in clinical measures. , 1992, BMJ.

[21]  Jacob Cohen A Coefficient of Agreement for Nominal Scales , 1960 .

[22]  H. Svanholm,et al.  Reproducibility of histomorphologic diagnoses with special reference to the kappa statistic , 1989, APMIS : acta pathologica, microbiologica, et immunologica Scandinavica.

[23]  J. Christensen,et al.  Pig health and production surveillance in Denmark: sampling design, data recording, and measures of disease frequency , 1994 .

[24]  A. Plint,et al.  Dating of bruises in children: an assessment of physician accuracy. , 2003, Pediatrics.

[25]  W. Grove Statistical Methods for Rates and Proportions, 2nd ed , 1981 .

[26]  B. Bonnett,et al.  Navicular Syndrome: Preliminary Assessment of Radiographic Scoring , 1995, Veterinary and Comparative Orthopaedics and Traumatology.

[27]  Jacob Cohen,et al.  Weighted kappa: Nominal scale agreement provision for scaled disagreement or partial credit. , 1968 .

[28]  A. Feinstein,et al.  High agreement but low kappa: II. Resolving the paradoxes. , 1990, Journal of clinical epidemiology.

[29]  R Williams,et al.  Inter-observer variation of symptoms and signs in jaundice. , 2008, Liver.

[30]  Graham Dunn,et al.  Review papers : Design and analysis of reliability studies , 1992 .

[31]  M. Heinonen,et al.  Diagnoses and Treatments in Health-Classified Fattening Herds Rearing Pigs All In – All Out , 2001, Acta veterinaria Scandinavica.

[32]  S. Gore Assessing clinical trials--between-observer variation. , 1981, British medical journal.

[33]  Bryan K. Smith,et al.  Evaluation of mild lameness in horses trotting on a treadmill by clinicians and interns or residents and correlation of their assessments with kinematic gait analysis. , 1998, American journal of veterinary research.

[34]  S. Edwards,et al.  Acute phase proteins in cattle: discrimination between acute and chronic inflammation , 1999, Veterinary Record.