How to measure the quality of the OSCE: A review of metrics – AMEE guide no. 49

With an increasing use of criterion-based assessment techniques in both undergraduate and postgraduate healthcare programmes, there is a consequent need to ensure the quality and rigour of these assessments. The obvious question for those responsible for delivering assessment is how is this ‘quality’ measured, and what mechanisms might there be that allow improvements in assessment quality over time to be demonstrated? Whilst a small base of literature exists, few papers give more than one or two metrics as measures of quality in Objective Structured Clinical Examinations (OSCEs). In this guide, aimed at assessment practitioners, the authors aim to review the metrics that are available for measuring quality and indicate how a rounded picture of OSCE assessment quality may be constructed by using a variety of such measures, and also to consider which characteristics of the OSCE are appropriately judged by which measure(s). The authors will discuss the quality issues both at the individual station level and across the complete clinical assessment as a whole, using a series of ‘worked examples’ drawn from OSCE data sets from the authors’ institution.

[1]  Brian Jolly,et al.  Assuring the quality of high-stakes undergraduate assessments of clinical competence , 2006, Medical teacher.

[2]  Geraldine Lear,et al.  Standard setting. , 2003, Nursing standard (Royal College of Nursing (Great Britain) : 1987).

[3]  Godfrey Pell,et al.  Setting standards for student assessment , 2006 .

[4]  M. Cusimano Standard setting in medical education , 1996, Academic medicine : journal of the Association of American Medical Colleges.

[5]  M. Homer,et al.  The impact of the inclusion of simulated patient ratings on the reliability of OSCE assessments under the borderline regression method , 2009, Medical teacher.

[6]  D. Streiner,et al.  Health measurement scales , 2008 .

[7]  A. Field Discovering statistics using SPSS for Windows. , 2000 .

[8]  R. Reznick,et al.  Comparing the psychometric properties of checklists and global rating scales for assessing performance on an OSCE‐format examination , 1998, Academic medicine : journal of the Association of American Medical Colleges.

[9]  T. Roberts,et al.  Assessor training: its effects on criterion‐based assessment in a medical context , 2008 .

[10]  J. Norcini,et al.  Setting standards on educational tests , 2003, Medical education.

[11]  J. Norcini,et al.  Workplace-based assessment as an educational tool: AMEE Guide No. 31 , 2007, Medical teacher.

[12]  C. V. D. van der Vleuten,et al.  Composite undergraduate clinical examinations: how should the components be combined to maximize reliability? , 2001, Medical education.

[13]  Vanessa C. Burch,et al.  Workplace-based assessment as an educational tool , 2007 .

[14]  David Newble,et al.  Techniques for measuring clinical competence: objective structured clinical examinations , 2004, Medical education.

[15]  Eric S. Holmboe,et al.  Faculty and the Observation of Trainees’ Clinical Skills: Problems and Opportunities , 2004, Academic medicine : journal of the Association of American Medical Colleges.

[16]  J. Colliver,et al.  A large-scale study of the reliabilities of checklist scores and ratings of interpersonal and communication skills evaluated on a standardized-patient examination , 1996, Advances in health sciences education : theory and practice.

[17]  J. Stevens Applied Multivariate Statistics for the Social Sciences , 1986 .

[18]  T. Swanwick,et al.  Workplace-based assessment. , 2009, British journal of hospital medicine.

[19]  K. Eva,et al.  An admissions OSCE: the multiple mini‐interview , 2004, Medical education.