Validity evidence for the Fundamentals of Laparoscopic Surgery (FLS) program as an assessment tool: a systematic review

BackgroundThe Fundamentals of Laparoscopic Surgery (FLS) program uses five simulation stations (peg transfer, precision cutting, loop ligation, and suturing with extracorporeal and intracorporeal knot tying) to teach and assess laparoscopic surgery skills. We sought to summarize evidence regarding the validity of scores from the FLS assessment.MethodsWe systematically searched for studies evaluating the FLS as an assessment tool (last search update February 26, 2013). We classified validity evidence using the currently standard validity framework (content, response process, internal structure, relations with other variables, and consequences).ResultsFrom a pool of 11,628 studies, we identified 23 studies reporting validity evidence for FLS scores. Studies involved residents (n = 19), practicing physicians (n = 17), and medical students (n = 8), in specialties of general (n = 17), gynecologic (n = 4), urologic (n = 1), and veterinary (n = 1) surgery. Evidence was most common in the form of relations with other variables (n = 22, most often expert–novice differences). Only three studies reported internal structure evidence (inter-rater or inter-station reliability), two studies reported content evidence (i.e., derivation of assessment elements), and three studies reported consequences evidence (definition of pass/fail thresholds). Evidence nearly always supported the validity of FLS total scores. However, the loop ligation task lacks discriminatory ability.ConclusionValidity evidence confirms expected relations with other variables and acceptable inter-rater reliability, but other validity evidence is sparse. Given the high-stakes use of this assessment (required for board eligibility), we suggest that more validity evidence is required, especially to support its content (selection of tasks and scoring rubric) and the consequences (favorable and unfavorable impact) of assessment.

[1]  G. Fried,et al.  Comparison of laparoscopic performance in vivo with performance measured in a laparoscopic simulator , 1999, Surgical Endoscopy.

[2]  Melina C Vassiliou,et al.  FLS and FES: comprehensive models of training and assessment. , 2010, The Surgical clinics of North America.

[3]  Melina C Vassiliou,et al.  A global assessment tool for evaluation of intraoperative laparoscopic skills. , 2005, American journal of surgery.

[4]  M. Roizen,et al.  Technology-enhanced simulation for health professions education: a systematic review and meta-analysis , 2012 .

[5]  David A Cook,et al.  Much ado about differences: why expert-novice comparisons add little to the validity argument , 2015, Advances in health sciences education : theory and practice.

[6]  Dimitrios Stefanidis,et al.  Higher Mental Workload is Associated With Poorer Laparoscopic Performance as Measured by the NASA-TLX Tool , 2010, Simulation in healthcare : journal of the Society for Simulation in Healthcare.

[7]  G. Fried,et al.  Proving the Value of Simulation in Laparoscopic Surgery , 2004, Annals of surgery.

[8]  G. Norman,et al.  Generalizability theory for the perplexed: A practical introduction and guide: AMEE Guide No. 68 , 2012, Medical teacher.

[9]  David A Cook,et al.  Technology-Enhanced Simulation to Assess Health Professionals: A Systematic Review of Validity Evidence, Research Methods, and Reporting Quality , 2013, Academic medicine : journal of the Association of American Medical Colleges.

[10]  Adeline M. Deladisma,et al.  The general surgery milestone project. , 2014, Journal of graduate medical education.

[11]  B. Dunkin Fundamentals of Endoscopic Surgery , 2012 .

[12]  G. Fried,et al.  Validity of the MISTELS simulator for laparoscopy training in urology. , 2005, Journal of endourology.

[13]  James R Korndorffer,et al.  Redefining simulator proficiency using automaticity theory. , 2007, American journal of surgery.

[14]  E. M. Ritter,et al.  Concurrent validity of augmented reality metrics applied to the fundamentals of laparoscopic surgery (FLS) , 2007, Surgical Endoscopy.

[15]  R. Wolterbeek,et al.  Laparoscopic skills simulator: construct validity and establishment of performance standards for residency training , 2008, Gynecological Surgery.

[16]  S. Hamstra,et al.  What counts as validity evidence? Examples and prevalence in a systematic review of simulation-based assessment , 2014, Advances in health sciences education : theory and practice.

[17]  G. Fried,et al.  Development of a model for training and evaluation of laparoscopic skills. , 1998, American journal of surgery.

[18]  C. Cao,et al.  Comparison of the sensitivity of physical and virtual laparoscopic surgical training simulators to the user’s level of experience , 2005, Surgical Endoscopy And Other Interventional Techniques.

[19]  Adam Dubrowski,et al.  Simulation in laparoscopic surgery: a concurrent validity study for FLS , 2008, Surgical Endoscopy.

[20]  I. Philibert,et al.  The next GME accreditation system--rationale and benefits. , 2012, The New England journal of medicine.

[21]  Harrith Hasson,et al.  Physical Reality Simulation for Training of Laparoscopists in the 21st Century. A Multispecialty, Multi-institutional Study , 2005, JSLS : Journal of the Society of Laparoendoscopic Surgeons.

[22]  G. Fried,et al.  Relationship between objective assessment of technical skills and subjective in-training evaluations in surgical residents. , 2004, Journal of the American College of Surgeons.

[23]  Monica Farcas,et al.  Development and validation of a pediatric laparoscopic surgery simulator. , 2011, Journal of pediatric surgery.

[24]  D. Moher,et al.  Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement , 2009, BMJ.

[25]  Harrith Hasson,et al.  Choosing the right physical laparoscopic simulator? Comparison of LTS2000-ISM60 with MISTELS: validation, correlation, and user satisfaction. , 2009, American journal of surgery.

[26]  S. Downing Validity: on the meaningful interpretation of assessment data , 2003, Medical education.

[27]  David A Cook,et al.  Linking Simulation-Based Educational Assessments and Patient-Related Outcomes: A Systematic Review and Meta-Analysis , 2015, Academic medicine : journal of the Association of American Medical Colleges.

[28]  R. J. Valentine,et al.  The general surgery milestones project. , 2014, Journal of the American College of Surgeons.

[29]  A. L. Trejos,et al.  Toward construct validity for a novel sensorized instrument-based minimally invasive surgery simulation system , 2011, Surgical Endoscopy.

[30]  Lee L Swanstrom,et al.  Beta test results of a new system assessing competence in laparoscopic surgery. , 2006, Journal of the American College of Surgeons.

[31]  D. Cook,et al.  Current concepts in validity and reliability for psychometric instruments: theory and application. , 2006, The American journal of medicine.

[32]  L. Way,et al.  Fundamentals of laparoscopic surgery , 1995 .

[33]  P. Bossuyt,et al.  Empirical evidence of design-related bias in studies of diagnostic tests. , 1999, JAMA.

[34]  Dimitrios Stefanidis,et al.  Robotic suturing on the FLS model possesses construct validity, is less physically demanding, and is favored by more surgeons compared with laparoscopy , 2011, Surgical Endoscopy.

[35]  J. R. Landis,et al.  The measurement of observer agreement for categorical data. , 1977, Biometrics.

[36]  A. Darzi,et al.  Observational tools for assessment of procedural skills: a systematic review. , 2011, American journal of surgery.

[37]  Alija Kulenović,et al.  Standards for Educational and Psychological Testing , 1999 .

[38]  G. Fried,et al.  The MISTELS program to measure technical skill in laparoscopic surgery , 2006, Surgical Endoscopy And Other Interventional Techniques.

[39]  V. Pankratz,et al.  Internal structure of mini-CEX scores for internal medicine residents: factor analysis and generalizability , 2010, Advances in health sciences education : theory and practice.

[40]  David A Cook,et al.  Association between funding and quality of published medical education research. , 2007, JAMA.

[41]  G. Fried,et al.  Evaluating laparoscopic skills: setting the pass/fail score for the MISTELS system. , 2003, Surgical endoscopy.

[42]  Suvranu De,et al.  Preliminary face and construct validation study of a virtual basic laparoscopic skill trainer. , 2010, Journal of laparoendoscopic & advanced surgical techniques. Part A.

[43]  G. Fried,et al.  FLS simulator performance predicts intraoperative laparoscopic skill , 2007, Surgical Endoscopy.

[44]  L. Swanström,et al.  Validity of using Fundamentals of Laparoscopic Surgery (FLS) program to assess laparoscopic competence for gynecologists , 2009, Surgical Endoscopy.

[45]  B. Fransson,et al.  Assessment of laparoscopic skills before and after simulation training with a canine abdominal model. , 2010, Journal of the American Veterinary Medical Association.