Assessing the surgical skills of trainees in the operating theatre: a prospective observational study of the methodology.

OBJECTIVES To compare user satisfaction and acceptability, reliability and validity of three different methods of assessing the surgical skills of trainees by direct observation in the operating theatre across a range of different surgical specialties and index procedures. DESIGN AND SETTING A 2-year prospective, observational study in the operating theatres of three teaching hospitals in Sheffield. METHODS The assessment methods were procedure-based assessment (PBA), Objective Structured Assessment of Technical Skills (OSATS) and Non-technical Skills for Surgeons (NOTSS). The specialties were obstetrics and gynaecology (O&G) and upper gastrointestinal, colorectal, cardiac, vascular and orthopaedic surgery. Two to four typical index procedures were selected from each specialty. Surgical trainees were directly observed performing typical index procedures and assessed using a combination of two of the three methods (OSATS or PBA and NOTSS for O&G, PBA and NOTSS for the other specialties) by the consultant clinical supervisor for the case and the anaesthetist and/or scrub nurse, as well as one or more independent assessors from the research team. OUTCOME MEASURES Information on user satisfaction and acceptability of each assessment method from both assessor and trainee perspectives was obtained from structured questionnaires. The reliability of each method was measured using generalisability theory. Aspects of validity included the internal structure of each tool and correlation between tools, construct validity, predictive validity, interprocedural differences, the effect of assessor designation and the effect of assessment on performance. RESULTS Of the 558 patients who were consented, a total of 437 (78%) cases were included in the study: 51 consultant clinical supervisors, 56 anaesthetists, 39 nurses, 2 surgical care practitioners and 4 independent assessors provided 1635 assessments on 85 trainees undertaking the 437 cases. A total of 749 PBAs, 695 NOTSS and 191 OSATSs were performed. Non-O&G clinical supervisors and trainees provided mixed, but predominantly positive, responses about a range of applications of PBA. Most felt that PBA was important in surgical education, and would use it again in the future and did not feel that it added time to the operating list. The overall satisfaction of O&G clinical supervisors and trainees with OSATS was not as high, and a majority of those who used both preferred PBA. A majority of anaesthetists and nurses felt that NOTSS allowed them to rate interpersonal skills (communication, teamwork and leadership) more easily than cognitive skills (situation awareness and decision-making), that it had formative value and that it was a valuable adjunct to the assessment of technical skills. PBA demonstrated high reliability (G > 0.8 for only three assessor judgements on the same index procedure). OSATS had lower reliability (G > 0.8 for five assessor judgements on the same index procedure). Both were less reliable on a mix of procedures because of strong procedure-specific factors. A direct comparison of PBA between O&G and non-O&G cases showed a striking difference in reliability. Within O&G, a good level of reliability (G > 0.8) could not be obtained using a feasible number of assessments. Conversely, the reliability within non-O&G cases was exceptionally high, with only two assessor judgements being required. The reasons for this difference probably include the more summative purpose of assessment in O&G and the much higher proportion of O&G trainees in this study with training concerns (42% vs 4%). The reliability of NOTSS was lower than that for PBA. Reliability for the same procedure (G > 0.8) required six assessor judgements. However, as procedure-specific factors exerted a lesser influence on NOTSS, reliability on a mix of procedures could be achieved using only eight assessor judgements. NOTSS also demonstrated a valid internal structure. The strongest correlations between NOTSS and PBA or OSATS were in the 'decision-making' domain. PBA and NOTSS showed better construct validity than OSATS, the year of training and the number of recent index procedures performed being significant independent predictors of performance. There was little variation in scoring between different procedures or different designations of assessor. CONCLUSIONS The results suggest that PBA is a reliable and acceptable method of assessing surgical skills, with good construct validity. Specialties that use OSATS may wish to consider changing the design or switching to PBA. Whatever workplace-based assessment method is used, the purpose, timing and frequency of assessment require detailed guidance. NOTSS is a promising tool for the assessment of non-technical skills, and surgical specialties may wish to consider its inclusion in their assessment framework. Further research is required into the use of health-care professionals other than consultant surgeons to assess trainees, the relationship between performance and experience, the educational impact of assessment and the additional value of video recording.

[1]  S. Roff,et al.  Development and validation of an instrument to measure the postgraduate clinical learning and teaching educational environment for hospital-based junior doctors in the UK , 2005, Medical teacher.

[2]  Richard Hays,et al.  In‐training assessment in postgraduate training for general practice , 1998, Medical education.

[3]  T S Murray,et al.  Use of videotaped consultations in summative assessment of trainees in general practice. , 1995, The British journal of general practice : the journal of the Royal College of General Practitioners.

[4]  A. G. Gallagher,et al.  The effect of escalating feedback on the acquisition of psychomotor skills for laparoscopy , 2007, Surgical Endoscopy.

[5]  R. Reznick,et al.  Verbal feedback from an expert is more effective than self-accessed feedback about motion efficiency in learning new surgical skills. , 2007, American journal of surgery.

[6]  M. Shapiro,et al.  High-Fidelity Medical Simulation as an Assessment Tool for Pediatric Residents' Airway Management Skills , 2007, Pediatric emergency care.

[7]  D. Newble,et al.  Assessing the technical skills of surgical trainees , 2005, The British journal of surgery.

[8]  D. Prytherch,et al.  A model for national outcome audit in vascular surgery. , 2001, European journal of vascular and endovascular surgery : the official journal of the European Society for Vascular Surgery.

[9]  R. Flin,et al.  Non-technical skills for surgeons in the operating room: a review of the literature. , 2006, Surgery.

[10]  D. Newble,et al.  The effect of assessments and examinations on the learning of medical students , 1983, Medical education.

[11]  Simon Paterson-Brown,et al.  Accuracy of medical staff assessment of trainees’ operative performance , 2005, Medical teacher.

[12]  M. Hewson,et al.  Giving feedback in medical education: verification of recommended techniques. , 1998, Journal of general internal medicine.

[13]  Ara Darzi,et al.  Comparison of bench test evaluations of surgical skill with live operating performance assessments. , 2004, Journal of the American College of Surgeons.

[14]  H. Davies,et al.  Improving the quality of outpatient clinic letters using the Sheffield Assessment Instrument for Letters (SAIL) , 2004, Medical education.

[15]  B. Goff,et al.  A six-year study of surgical teaching and skills evaluation for obstetric/gynecologic residents in porcine and inanimate surgical models. , 2005, American journal of obstetrics and gynecology.

[16]  C. Muller,et al.  Effects of a laboratory-based skills curriculum on laparoscopic proficiency: a randomized trial. , 2002, American journal of obstetrics and gynecology.

[17]  R. Reznick,et al.  Comparing the psychometric properties of checklists and global rating scales for assessing performance on an OSCE‐format examination , 1998, Academic medicine : journal of the Association of American Medical Colleges.

[18]  B Jolly,et al.  The assessment of poorly performing doctors: the development of the assessment programmes for the General Medical Council’s Performance Procedures , 2001, Medical education.

[19]  E. Copeland,et al.  A prospective study of patient safety in the operating room , 2007 .

[20]  R. Reznick,et al.  Teaching and testing technical skills. , 1993, American journal of surgery.

[21]  C. Warlow,et al.  MRC European Carotid Surgery Trial: interim results for symptomatic patients with severe (70-99%) or with mild (0-29%) carotid stenosis , 1991, The Lancet.

[22]  S M Williams,et al.  Quality control: an application of the cusum. , 1992, BMJ.

[23]  Brian Jolly,et al.  Assessing health professionals , 2002, Medical education.

[24]  J. Carline,et al.  Use of peer ratings to evaluate physician performance. , 1993, JAMA.

[25]  E. Frezza,et al.  Competency-based instruction in critical invasive skills improves both resident performance and patient safety. , 1998, Surgery.

[26]  Steven M Downing,et al.  Reliability: on the reproducibility of assessment data , 2004, Medical education.

[27]  Ronald M. Harden,et al.  AMEE Guide No. 14: Outcome-based education: Part 1-An introduction to outcome-based education , 1999 .

[28]  Rhona Flin,et al.  Experience matters: comparing novice and expert ratings of non‐technical skills using the NOTSS system , 2009, ANZ journal of surgery.

[29]  G. Norman The long case versus objective structured clinical examinations , 2002, BMJ : British Medical Journal.

[30]  C. Roberts,et al.  ‘I’m pickin' up good regressions': the governance of generalisability analyses , 2007, Medical education.

[31]  R. Reznick,et al.  Objective structured assessment of technical skill (OSATS) for surgical residents , 1997, The British journal of surgery.

[32]  G. Miller The assessment of clinical skills/competence/performance , 1990, Academic medicine : journal of the Association of American Medical Colleges.

[33]  The Good Assessment Guide: A Practical Guide to Assessment and Appraisal for Higher Specialist Training , 1997 .

[34]  E. Thorndike A constant error in psychological ratings. , 1920 .

[35]  L. Shulman,et al.  Medical problem-solving. , 2014, Journal of medical education.

[36]  A. Fordham Using a competency based approach in nurse education. , 2005, Nursing standard (Royal College of Nursing (Great Britain) : 1987).

[37]  C. Galasko,et al.  Unsupervised surgical training , 1997, British medical journal.

[38]  R. Reznick,et al.  Reliability and construct validity of a structured technical skills assessment form. , 1994, American journal of surgery.

[39]  Brian Jolly,et al.  Generalisability: a key to unlock professional assessment , 2002, Medical education.

[40]  R. Flin,et al.  Anaesthetists' Non-Technical Skills (ANTS): evaluation of a behavioural marker system. , 2003, British journal of anaesthesia.

[41]  C. V. D. van der Vleuten,et al.  Assessing professional competence : from methods to programmes , 2005 .

[42]  David M Studdert,et al.  Analysis of errors reported by surgeons at three teaching hospitals. , 2003, Surgery.

[43]  K. Cassar Development of an instrument to measure the surgical operating theatre learning environment as perceived by basic surgical trainees , 2004, Medical teacher.

[44]  H C FALK,et al.  The training of a surgeon. , 1958, The Journal of the Albert Einstein Medical Center, Philadelphia.

[45]  A. Ziv,et al.  Features and uses of high-fidelity medical simulations that lead to effective learning: a BEME systematic review , 2005, Medical teacher.

[46]  P. Bassett,et al.  Simulator training improves practical skills in therapeutic GI endoscopy: results from a randomized, blinded, controlled study. , 2009, Gastrointestinal endoscopy.

[47]  C. V. D. van der Vleuten,et al.  The assessment of professional competence: Developments, research and practical implications. , 1996, Advances in health sciences education : theory and practice.

[48]  T Pickersgill,et al.  The European working time directive for doctors in training , 2001, BMJ : British Medical Journal.

[49]  A. V. Van rij,et al.  Cusum as an aid to early assessment of the surgical trainee , 1995, The British journal of surgery.

[50]  Teodor P Grantcharov,et al.  Learning curves and impact of previous operative experience on performance on a virtual reality simulator to test laparoscopic surgical skills. , 2003, American journal of surgery.

[51]  I. Dickinson,et al.  Guide to the Assessment of Competence and Performance in Practising Surgeons , 2009, ANZ journal of surgery.

[52]  D. Upton,et al.  Surgeons' experience of learning psychological skills: a preliminary evaluation of a psychological skills training course. , 2009, Annals of the Royal College of Surgeons of England.

[53]  Cairncross,et al.  Reforming higher specialist training in the United Kingdom – a step along the continuum of medical education , 1999, Medical education.

[54]  T. Grantcharov,et al.  Randomized clinical trial of virtual reality simulation for laparoscopic skills training , 2004, The British journal of surgery.

[55]  D. Newble,et al.  Developing an outcome‐focused core curriculum , 2005, Medical education.

[56]  A. Wolf Competence-Based Assessment , 1995 .

[57]  J. Birkmeyer,et al.  Simulation Improves Resident Performance in Catheter-Based Intervention: Results of a Randomized, Controlled Study , 2006, Annals of surgery.

[58]  F Sturmans,et al.  Does competence of general practitioners predict their performance? Comparison between examination setting and actual practice. , 1991, BMJ.

[59]  J. Rethans,et al.  The relationship between competence and performance: implications for assessing practice performance , 2002, Medical education.

[60]  D. Rowley,et al.  Assessment of performance in orthopaedic training. , 2005, The Journal of bone and joint surgery. British volume.

[61]  J. Rethans,et al.  Assessment of general practitioners by video observation of communicative and medical performance in daily practice: issues of validity, reliability and feasibility , 1999, Medical education.

[62]  J. Rethans,et al.  Assessment in general practice: the predictive value of written‐knowledge tests and a multiple‐station examination for actual medical performance in daily practice , 1999, Medical education.

[63]  G. Elwyn,et al.  Doctor performance assessment in daily practise: does it help doctors or not? A systematic review , 2007, Medical education.

[64]  Rhona Flin,et al.  Debriefing surgeons on non-technical skills (NOTSS) , 2008, Cognition, Technology & Work.

[65]  R. Rege,et al.  Measuring operative performance after laparoscopic skills training: edited videotape versus direct observation. , 2000, Journal of laparoendoscopic & advanced surgical techniques. Part A.

[66]  J. Wilting,et al.  The freedom to learn. , 1974, Nursing papers. Perspectives en nursing.

[67]  Benjamin S. Bloom,et al.  Taxonomy of Educational Objectives: The Classification of Educational Goals. , 1957 .

[68]  J. Collins Clinical supervision of SpRs: where does it happen, when does it happen and is it effective? , 2003, Medical education.

[69]  R Flin,et al.  Development of a rating system for surgeons' non‐technical skills , 2006, Medical education.

[70]  Larry Ambrose,et al.  The power of feedback. , 2002, Healthcare executive.

[71]  J. Beard,et al.  Developing an education and assessment framework for the Foundation Programme , 2005, Medical education.

[72]  J. Beard Assessment of surgical skills of trainees in the UK. , 2008, Annals of the Royal College of Surgeons of England.

[73]  R Reznick,et al.  Using operative outcome to assess technical skill. , 2000, American journal of surgery.

[74]  L. Jacoby,et al.  Knowledge and clinical problem‐solving , 1985, Medical education.

[75]  J A Reekers,et al.  The role of simulation in training endovascular interventions. , 2008, European journal of vascular and endovascular surgery : the official journal of the European Society for Vascular Surgery.

[76]  Geoffrey R. Norman,et al.  Performance-Based Assessment: Lessons From the Health Professions , 1995 .

[77]  M. Philbin,et al.  Positive effects of a clinical performance assessment program , 1991, Academic medicine : journal of the Association of American Medical Colleges.

[78]  Donovan A. McFarlane,et al.  Evaluating Training Programs: The Four Levels , 2006 .

[79]  D. Risucci,et al.  Ratings of surgical residents by self, supervisors and peers. , 1989, Surgery, gynecology & obstetrics.

[80]  R. Gold Roles in Sociological Field Observations , 1958 .

[81]  S. Dawson,et al.  Procedural simulation: a primer. , 2006, Journal of vascular and interventional radiology : JVIR.

[82]  Questionnaire construction and question writing for research in medical education. , 1988, Medical education.

[83]  A. Spigelman Consultant surgeons' opinion of the skills required of basic surgical trainees , 2000, The British journal of surgery.

[84]  D. Fenner,et al.  Surgical skills assessment: a blinded examination of obstetrics and gynecology residents. , 2002, American journal of obstetrics and gynecology.

[85]  J. Norcini,et al.  Workplace-based assessment as an educational tool: AMEE Guide No. 31 , 2007, Medical teacher.

[86]  J. Morrison,et al.  ABC of learning and teaching in medicine: Evaluation. , 2003, BMJ.

[87]  R. Satava,et al.  Virtual reality as a metric for the assessment of laparoscopic psychomotor skills , 2002, Surgical Endoscopy And Other Interventional Techniques.

[88]  David J. Woehr,et al.  Rater training for performance appraisal: A quantitative review , 1994 .

[89]  Rhona Flin,et al.  Surgeons’ Non-technical Skills in the Operating Room: Reliability Testing of the NOTSS Behavior Rating System , 2008, World Journal of Surgery.

[90]  Jessica L. Smith,et al.  High-fidelity medical simulation as an assessment tool for radiology residents' acute contrast reaction management skills. , 2009, Journal of the American College of Radiology : JACR.

[91]  R. Aggarwal,et al.  Objective assessment of gynecologic laparoscopic skills using the LapSimGyn virtual reality simulator , 2006, Surgical Endoscopy And Other Interventional Techniques.

[92]  L W T Schuwirth,et al.  When enough is enough: a conceptual basis for fair and defensible practice performance assessment , 2002, Medical education.

[93]  Daniel B. Jones,et al.  Laparoscopic training on bench models: better and more cost effective than operating room experience? , 2000, Journal of the American College of Surgeons.

[94]  A. Darzi,et al.  Objective assessment of technical skills in surgery , 2003, BMJ : British Medical Journal.

[95]  J. Beard,et al.  Setting standards for the assessment of operative competence. , 2005, European journal of vascular and endovascular surgery : the official journal of the European Society for Vascular Surgery.