The reliability and validity of a portfolio designed as a programmatic assessment of performance in an integrated clinical placement

BackgroundLittle is known about the technical adequacy of portfolios in reporting multiple complex academic and performance-based assessments. We explored, first, the influencing factors on the precision of scoring within a programmatic assessment of student learning outcomes within an integrated clinical placement. Second, the degree to which validity evidence supported interpretation of student scores.MethodsWithin generalisability theory, we estimated the contribution that each wanted factor (i.e. student capability) and unwanted factors (e.g. the impact of assessors) made to the variation in portfolio task scores. Relative and absolute standard errors of measurement provided a confidence interval around a pre-determined pass/fail standard for all six tasks. Validity evidence was sought through demonstrating the internal consistency of the portfolio and exploring the relationship of student scores with clinical experience.ResultsThe mean portfolio mark for 257 students, across 372 raters, based on six tasks, was 75.56 (SD, 6.68). For a single student on one assessment task, 11% of the variance in scores was due to true differences in student capability. The most significant interaction was context specificity (49%), the tendency for one student to engage with one task and not engage with another task. Rater subjectivity was 29%. An absolute standard error of measurement of 4.74%, gave a 95% CI of +/- 9.30%, and a 68% CI of +/- 4.74% around a pass/fail score of 57%. Construct validity was supported by demonstration of an assessment framework, the internal consistency of the portfolio tasks, and higher scores for students who did the clinical placement later in the academic year.ConclusionA portfolio designed as a programmatic assessment of an integrated clinical placement has sufficient evidence of validity to support a specific interpretation of student scores around passing a clinical placement. It has modest precision in assessing students’ achievement of a competency standard. There were identifiable areas for reducing measurement error and providing more certainty around decision-making. Reducing the measurement error would require engaging with the student body on the value of the tasks, more focussed academic and clinical supervisor training, and revisiting the rubric of the assessment in the light of feedback.

[1]  David Boud,et al.  Reframing assessment as if learning were important , 2007 .

[2]  Descriptors Educational,et al.  of Educational Measurement , 1988 .

[3]  C. Roberts,et al.  ‘I’m pickin' up good regressions': the governance of generalisability analyses , 2007, Medical education.

[4]  David Newble,et al.  Techniques for measuring clinical competence: objective structured clinical examinations , 2004, Medical education.

[5]  BMC Medical Education , 2006 .

[6]  F M Farrar,et al.  ROYAL AUSTRALIAN COLLEGE OF GENERAL PRACTITIONERS , 1973, Canadian family physician Medecin de famille canadien.

[7]  C. V. D. van der Vleuten,et al.  A model for programmatic assessment fit for purpose , 2012, Medical teacher.

[8]  J Murtagh Common problems: a safe diagnostic strategy. , 1990, Australian family physician.

[9]  G. Norman,et al.  How specific is case specificity? , 2006, Medical education.

[10]  Timothy J. Wood,et al.  Exploring the role of first impressions in rater-based assessments , 2014, Advances in health sciences education : theory and practice.

[11]  Examiner perceptions of a portfolio assessment process , 2010, Medical teacher.

[12]  A. Nichols,et al.  Australian College of Rural and Remote Medicine. , 2002, The Australian journal of rural health.

[13]  David Snadden Mary Thomas,et al.  The use of portfolio learning in medical education , 1998 .

[14]  Cees P. M. van der Vleuten,et al.  Programmatic assessment and Kane's validity perspective. , 2012 .

[15]  A. Scherpbier,et al.  Workplace learning from a socio-cultural perspective: creating developmental space during the general practice clerkship , 2010, Advances in health sciences education : theory and practice.

[16]  C. V. D. van der Vleuten,et al.  The assessment of professional competence: Developments, research and practical implications. , 1996, Advances in health sciences education : theory and practice.

[17]  C. Rees The use (and abuse) of the term ‘portfolio’ , 2005, Medical education.

[18]  C. V. D. van der Vleuten,et al.  Programmatic assessment and Kane’s validity perspective , 2012, Medical education.

[19]  George D. Kuh,et al.  Student Engagement and Student Learning: Testing the Linkages* , 2006 .

[20]  C. Popovic,et al.  The educational effects of portfolios on undergraduate student learning: A Best Evidence Medical Education (BEME) systematic review. BEME Guide No. 11 , 2009, Medical teacher.

[21]  S. Downing Validity: on the meaningful interpretation of assessment data , 2003, Medical education.

[22]  C. V. D. van der Vleuten,et al.  Composite undergraduate clinical examinations: how should the components be combined to maximize reliability? , 2001, Medical education.

[23]  A. Muijtjens,et al.  Workplace-based assessment: raters’ performance theories and constructs , 2012, Advances in Health Sciences Education.

[24]  C. Vleuten Programmatic assessment: From assessment of learning to assessment for learning , 2011 .

[25]  Erik W Driessen,et al.  Portfolios for assessment and learning: AMEE Guide no. 45 , 2009, Medical teacher.

[26]  Cees P M van der Vleuten,et al.  Conditions for successful reflective use of portfolios in undergraduate medical education , 2005, Medical education.

[27]  Brian Jolly,et al.  Assuring the quality of high-stakes undergraduate assessments of clinical competence , 2006, Medical teacher.

[28]  Edward H. Haertel,et al.  Generalizability Analysis for Performance Assessments of Student Achievement or School Effectiveness , 1997 .

[29]  J. Thistlethwaite,et al.  A review of longitudinal community and hospital placements in medical education: BEME Guide No. 26 , 2013, Medical teacher.

[30]  C. Roberts,et al.  Evaluating peer teaching about chronic disease , 2014, The clinical teacher.

[31]  C. Roberts,et al.  The reliability and validity of a matrix to assess the completed reflective personal development plans of general practitioners , 2006, Medical education.

[32]  Mark D. Reckase,et al.  Demonstration of Portfolios to Assess Competency of Residents , 2004, Advances in health sciences education : theory and practice.

[33]  C. Rees,et al.  The reliability of assessment criteria for undergraduate medical students' communication skills portfolios: the Nottingham experience , 2004, Medical education.

[34]  Eva Nick,et al.  The dependability of behavioral measurements: theory of generalizability for scores and profiles , 1973 .

[35]  Chris Roberts,et al.  Portfolio‐based assessments in medical education: are they valid and reliable for summative purposes? , 2002, Medical education.

[36]  Val Wass,et al.  Portfolios in medical education: why do they meet with mixed success? A systematic review , 2007, Medical education.

[37]  D. Boud,et al.  Rethinking assessment in higher education : learning for the longer term , 2007 .

[38]  Kevin W Eva,et al.  Exploring the impact of mental workload on rater-based assessments , 2012, Advances in Health Sciences Education.

[39]  H. Britt,et al.  Prevalence and patterns of multimorbidity in Australia , 2008, The Medical journal of Australia.

[40]  A. Muijtjens,et al.  Portfolio assessment during medical internships: How to obtain a reliable and feasible assessment procedure? , 2009, Education for health.

[41]  J. Norcini,et al.  Construct Validity of the MiniClinical Evaluation Exercise (MiniCEX) , 2003, Academic medicine : journal of the Association of American Medical Colleges.

[42]  R. Linn Educational measurement, 3rd ed. , 1989 .

[43]  C. V. D. van der Vleuten,et al.  Programmatic assessment: From assessment of learning to assessment for learning , 2011, Medical teacher.

[44]  J. Ker,et al.  Student perceptions of a portfolio assessment process , 2009, Medical education.

[45]  E. Dannefer Beyond assessment of learning toward assessment for learning: Educating tomorrow's physicians , 2013, Medical teacher.

[46]  N. Oswald,et al.  Evaluating primary care as a base for medical education: the report of the Cambridge Community‐based Clinical Course , 2001, Medical education.