Seeing the same thing differently

Assessors’ scores in performance assessments are known to be highly variable. Attempted improvements through training or rating format have achieved minimal gains. The mechanisms that contribute to variability in assessors’ scoring remain unclear. This study investigated these mechanisms. We used a qualitative approach to study assessors’ judgements whilst they observed common simulated videoed performances of junior doctors obtaining clinical histories. Assessors commented concurrently and retrospectively on performances, provided scores and follow-up interviews. Data were analysed using principles of grounded theory. We developed three themes that help to explain how variability arises: Differential Salience—assessors paid attention to (or valued) different aspects of the performances to different degrees; Criterion Uncertainty—assessors’ criteria were differently constructed, uncertain, and were influenced by recent exemplars; Information Integration—assessors described the valence of their comments in their own unique narrative terms, usually forming global impressions. Our results (whilst not precluding the operation of established biases) describe mechanisms by which assessors’ judgements become meaningfully-different or unique. Our results have theoretical relevance to understanding the formative educational messages that performance assessments provide. They give insight relevant to assessor training, assessors’ ability to be observationally “objective” and to the educational value of narrative comments (in contrast to numerical ratings).

[1]  Eric S Holmboe,et al.  Effects of Training in Direct Observation of Medical Residents' Clinical Competence , 2004, Annals of Internal Medicine.

[2]  A. Kluger,et al.  Feedback, the various tasks of the doctor, and the feedforward alternative , 2010, Medical education.

[3]  K. A. Ericsson,et al.  Verbal reports as data. , 1980 .

[4]  Angelo S. DeNisi,et al.  A cognitive view of the performance appraisal process: A model and research propositions , 1984 .

[5]  B. Clauser,et al.  Use of the Mini-Clinical Evaluation Exercise to Rate Examinee Performance on a Multiple-Station Clinical Skills Examination: A Validity Study , 2006, Academic medicine : journal of the Association of American Medical Colleges.

[6]  D. Streiner,et al.  Health measurement scales , 2008 .

[7]  Richard J. Klimoski,et al.  Accountability forces in performance appraisal , 1990 .

[8]  Jennifer Cleland,et al.  Identifying the factors that determine feedback given to undergraduate medical students following formative mini‐CEX assessments , 2007, Medical education.

[9]  D. Cook,et al.  Does scale length matter? A comparison of nine- versus five-point rating scales for the mini-CEX , 2009, Advances in health sciences education : theory and practice.

[10]  A. Wragg,et al.  Implementing workplace‐based assessment across the medical specialties in the United Kingdom , 2008, Medical education.

[11]  J. Norcini,et al.  Assessing the Comparability of Standardized Patient and Physician Evaluations of Clinical Skills , 2002, Advances in health sciences education : theory and practice.

[12]  S. Durning,et al.  Constructing a Validity Argument for the Mini-Clinical Evaluation Exercise: A Review of the Research , 2010, Academic medicine : journal of the Association of American Medical Colleges.

[13]  Deidra J. Schleicher,et al.  A new frame for frame-of-reference training: enhancing the construct validity of assessment centers. , 2002, The Journal of applied psychology.

[14]  H. Mandin,et al.  Can Concept Sorting Provide a Reliable, Valid and Sensitive Measure of Medical Knowledge Structure? , 2007, Advances in health sciences education : theory and practice.

[15]  Jack M. Feldman,et al.  Beyond Attribution Theory: Cognitive Processes in Performance Appraisal , 1981 .

[16]  D. Gelb,et al.  Education Research: Bias and poor interrater reliability in evaluating the neurology clinical skills examination , 2009, Neurology.

[17]  T. Chartrand,et al.  THE UNBEARABLE AUTOMATICITY OF BEING , 1999 .

[18]  J. Norcini,et al.  Examiner Differences in the Mini-Cex , 1997, Advances in health sciences education : theory and practice.

[19]  David Newble,et al.  Techniques for measuring clinical competence: objective structured clinical examinations , 2004, Medical education.

[20]  Jonathan Silverman,et al.  Marrying Content and Process in Clinical Method Teaching: Enhancing the Calgary–Cambridge Guides , 2003, Academic medicine : journal of the Association of American Medical Colleges.

[21]  Dawn Martin Martin's Map: a conceptual framework for teaching and learning the medical interview using a patient‐centred approach , 2003, Medical education.

[22]  A. Muijtjens,et al.  Workplace-based assessment: effects of rater expertise , 2010, Advances in health sciences education : theory and practice.

[23]  J. Norcini,et al.  Construct Validity of the MiniClinical Evaluation Exercise (MiniCEX) , 2003, Academic medicine : journal of the Association of American Medical Colleges.

[24]  I. Haq,et al.  Exploring the underperformance of male and minority ethnic medical students in first year clinical examinations , 2008, Advances in health sciences education : theory and practice.

[25]  Timothy D. Wilson,et al.  Telling more than we can know: Verbal reports on mental processes. , 1977 .

[26]  William C McGaghie,et al.  SPECIAL ARTICLE: Cognitive, Social and Environmental Sources of Bias in Clinical Performance Ratings , 2003, Teaching and learning in medicine.

[27]  David A. Cook,et al.  Effect of Rater Training on Reliability and Accuracy of Mini-CEX Scores: A Randomized, Controlled Trial , 2008, Journal of General Internal Medicine.

[28]  J. Kogan,et al.  Opening the black box of clinical skills assessment via observation: a conceptual model , 2011, Medical education.

[29]  S. Lurie,et al.  Measurement of the General Competencies of the Accreditation Council for Graduate Medical Education: A Systematic Review , 2009, Academic medicine : journal of the Association of American Medical Colleges.

[30]  I. M. Jawahar,et al.  WHERE ALL THE CHILDREN ARE ABOVE AVERAGE: THE PERFORMANCE APPRAISAL PURPOSE EFFECT , 1997 .

[31]  P. O'Callaghan Transient loss of consciousness , 2012 .

[32]  Henk G. Schmidt,et al.  On the Role of Biomedical Knowledge in Clinical Reasoning by Experts, Intermediates and Novices , 1992, Cogn. Sci..

[33]  C. V. D. van der Vleuten,et al.  In-training assessment using direct observation of single-patient encounters: a literature review , 2010, Advances in health sciences education : theory and practice.

[34]  J. Norcini The Mini Clinical Evaluation Exercise , 2005 .

[35]  J. Norcini,et al.  Work based assessment , 2003, BMJ : British Medical Journal.

[36]  Eric S. Holmboe,et al.  Tools for direct observation and assessment of clinical skills of medical trainees: a systematic review. , 2009, JAMA.

[37]  J. Kogan,et al.  What Drives Faculty Ratings of Residents' Clinical Skills? The Impact of Faculty's Own Clinical Skills , 2010, Academic medicine : journal of the Association of American Medical Colleges.

[38]  Timothy D. Wilson,et al.  The Proper Protocol: Validity and Completeness of Verbal Reports , 1994 .

[39]  David J. Woehr,et al.  Rater training for performance appraisal: A quantitative review , 1994 .

[40]  Anselm L. Strauss,et al.  Basics of qualitative research : techniques and procedures for developing grounded theory , 1998 .

[41]  Marjan J. B. Govaerts,et al.  Broadening Perspectives on Clinical Performance Assessment: Rethinking the Nature of In-training Assessment , 2007, Advances in health sciences education : theory and practice.

[42]  V. Pankratz,et al.  Internal structure of mini-CEX scores for internal medicine residents: factor analysis and generalizability , 2010, Advances in health sciences education : theory and practice.

[43]  S. Brutus Words versus numbers: A theoretical exploration of giving and receiving narrative comments in performance appraisal , 2010 .

[44]  Kevin W Eva,et al.  Rater-Based Assessments as Social Judgments: Rethinking the Etiology of Rater Errors , 2011, Academic medicine : journal of the Association of American Medical Colleges.

[45]  L. Pangaro,et al.  Evaluation of a novel assessment form for observing medical residents: a randomised, controlled trial , 2008, Medical education.

[46]  C. V. D. van der Vleuten,et al.  Validity, reliability, feasibility and satisfaction of the Mini-Clinical Evaluation Exercise (Mini-CEX) for cardiology residency training , 2007, Medical teacher.

[47]  J. Norcini The Mini Clinical Evaluation Exercise (mini‐CEX) , 2005 .

[48]  G. Regehr,et al.  Toward Authentic Clinical Evaluation: Pitfalls in the Pursuit of Competency , 2010, Academic medicine : journal of the Association of American Medical Colleges.

[49]  Cees P. M. van der Vleuten,et al.  Assessing professional competence: from methods to programmes , 2005 .

[50]  A. Merry,et al.  Mini-clinical evaluation exercise in anaesthesia training. , 2009, British journal of anaesthesia.

[51]  S. Thammasitboon,et al.  Exploring Individual Opinions of Potential Evaluators in a 360-Degree Assessment: Four Distinct Viewpoints of a Competent Resident , 2008, Teaching and learning in medicine.