Literature DB >> 26133652

A study of crowdsourced segment-level surgical skill assessment using pairwise rankings.

Anand Malpani1, S Swaroop Vedula, Chi Chiung Grace Chen, Gregory D Hager.   

Abstract

PURPOSE: Currently available methods for surgical skills assessment are either subjective or only provide global evaluations for the overall task. Such global evaluations do not inform trainees about where in the task they need to perform better. In this study, we investigated the reliability and validity of a framework to generate objective skill assessments for segments within a task, and compared assessments from our framework using crowdsourced segment ratings from surgically untrained individuals and expert surgeons against manually assigned global rating scores.
METHODS: Our framework includes (1) a binary classifier trained to generate preferences for pairs of task segments (i.e., given a pair of segments, specification of which one was performed better), (2) computing segment-level percentile scores based on the preferences, and (3) predicting task-level scores using the segment-level scores. We conducted a crowdsourcing user study to obtain manual preferences for segments within a suturing and knot-tying task from a crowd of surgically untrained individuals and a group of experts. We analyzed the inter-rater reliability of preferences obtained from the crowd and experts, and investigated the validity of task-level scores obtained using our framework. In addition, we compared accuracy of the crowd and expert preference classifiers, as well as the segment- and task-level scores obtained from the classifiers.
RESULTS: We observed moderate inter-rater reliability within the crowd (Fleiss' kappa, κ = 0.41) and experts (κ = 0.55). For both the crowd and experts, the accuracy of an automated classifier trained using all the task segments was above par as compared to the inter-rater agreement [crowd classifier 85 % (SE 2 %), expert classifier 89 % (SE 3 %)]. We predicted the overall global rating scores (GRS) for the task with a root-mean-squared error that was lower than one standard deviation of the ground-truth GRS. We observed a high correlation between segment-level scores (ρ ≥ 0.86) obtained using the crowd and expert preference classifiers. The task-level scores obtained using the crowd and expert preference classifier were also highly correlated with each other (ρ ≥ 0.84), and statistically equivalent within a margin of two points (for a score ranging from 6 to 30). Our analyses, however, did not demonstrate statistical significance in equivalence of accuracy between the crowd and expert classifiers within a 10 % margin.
CONCLUSIONS: Our framework implemented using crowdsourced pairwise comparisons leads to valid objective surgical skill assessment for segments within a task, and for the task overall. Crowdsourcing yields reliable pairwise comparisons of skill for segments within a task with high efficiency. Our framework may be deployed within surgical training programs for objective, automated, and standardized evaluation of technical skills.

Mesh:

Year:  2015        PMID: 26133652     DOI: 10.1007/s11548-015-1238-6

Source DB:  PubMed          Journal:  Int J Comput Assist Radiol Surg        ISSN: 1861-6410            Impact factor:   2.924


  21 in total

1.  The relationship between motion analysis and surgical technical assessments.

Authors:  Vivek Datta; Avril Chang; Sean Mackay; Ara Darzi
Journal:  Am J Surg       Date:  2002-07       Impact factor: 2.565

2.  Global evaluative assessment of robotic skills: validation of a clinical assessment tool to measure robotic surgical skills.

Authors:  Alvin C Goh; David W Goldfarb; James C Sander; Brian J Miles; Brian J Dunkin
Journal:  J Urol       Date:  2011-11-17       Impact factor: 7.450

3.  Synchronized video and motion analysis for the assessment of procedures in the operating theater.

Authors:  Aristotelis Dosis; Rajesh Aggarwal; Fernando Bello; Krishna Moorthy; Yaron Munz; Duncan Gillies; Ara Darzi
Journal:  Arch Surg       Date:  2005-03

4.  Randomized controlled trial on the effect of coaching in simulated laparoscopic training.

Authors:  Simon J Cole; Hugh Mackenzie; Joon Ha; George B Hanna; Danilo Miskovic
Journal:  Surg Endosc       Date:  2013-11-07       Impact factor: 4.584

5.  Why Johnny cannot operate.

Authors:  Richard H Bell
Journal:  Surgery       Date:  2009-10       Impact factor: 3.982

Review 6.  Resident education in 2011: three key challenges on the road ahead.

Authors:  Erik G Van Eaton; John L Tarpley; Carmen C Solorzano; Clifford S Cho; Sharon M Weber; Paula M Termuhlen
Journal:  Surgery       Date:  2011-02-05       Impact factor: 3.982

7.  Can masses of non-experts train highly accurate image classifiers? A crowdsourcing approach to instrument segmentation in laparoscopic images.

Authors:  Lena Maier-Hein; Sven Mersmann; Daniel Kondermann; Sebastian Bodenstedt; Alexandro Sanchez; Christian Stock; Hannes Gotz Kenngott; Mathias Eisenmann; Stefanie Speidel
Journal:  Med Image Comput Comput Assist Interv       Date:  2014

8.  Crowd-Sourced Assessment of Technical Skills: a novel method to evaluate surgical performance.

Authors:  Carolyn Chen; Lee White; Timothy Kowalewski; Rajesh Aggarwal; Chris Lintott; Bryan Comstock; Katie Kuksenok; Cecilia Aragon; Daniel Holst; Thomas Lendvay
Journal:  J Surg Res       Date:  2013-10-10       Impact factor: 2.192

9.  Markov modeling of minimally invasive surgery based on tool/tissue interaction and force/torque signatures for evaluating surgical skills.

Authors:  J Rosen; B Hannaford; C G Richards; M N Sinanan
Journal:  IEEE Trans Biomed Eng       Date:  2001-05       Impact factor: 4.538

10.  Objective measures for longitudinal assessment of robotic surgery training.

Authors:  Rajesh Kumar; Amod Jog; Balazs Vagvolgyi; Hiep Nguyen; Gregory Hager; Chi Chiung Grace Chen; David Yuh
Journal:  J Thorac Cardiovasc Surg       Date:  2011-12-14       Impact factor: 5.209

View more
  10 in total

1.  Video assessment of laparoscopic skills by novices and experts: implications for surgical education.

Authors:  Celine Yeung; Brian Carrillo; Victor Pope; Shahob Hosseinpour; J Ted Gerstle; Georges Azzie
Journal:  Surg Endosc       Date:  2017-02-15       Impact factor: 4.584

2.  A computer vision technique for automated assessment of surgical performance using surgeons' console-feed videos.

Authors:  Amir Baghdadi; Ahmed A Hussein; Youssef Ahmed; Lora A Cavuoto; Khurshid A Guru
Journal:  Int J Comput Assist Radiol Surg       Date:  2018-11-20       Impact factor: 2.924

3.  Query-by-example surgical activity detection.

Authors:  Yixin Gao; S Swaroop Vedula; Gyusung I Lee; Mija R Lee; Sanjeev Khudanpur; Gregory D Hager
Journal:  Int J Comput Assist Radiol Surg       Date:  2016-04-12       Impact factor: 2.924

4.  Meaningful Assessment of Robotic Surgical Style using the Wisdom of Crowds.

Authors:  M Ershad; R Rege; A Majewicz Fey
Journal:  Int J Comput Assist Radiol Surg       Date:  2018-03-24       Impact factor: 2.924

Review 5.  Objective Assessment of Surgical Technical Skill and Competency in the Operating Room.

Authors:  S Swaroop Vedula; Masaru Ishii; Gregory D Hager
Journal:  Annu Rev Biomed Eng       Date:  2017-03-27       Impact factor: 9.590

6.  Explaining a model predicting quality of surgical practice: a first presentation to and review by clinical experts.

Authors:  Arthur Derathé; Fabian Reche; Pierre Jannin; Alexandre Moreau-Gaudry; Bernard Gibaud; Sandrine Voros
Journal:  Int J Comput Assist Radiol Surg       Date:  2021-06-18       Impact factor: 2.924

Review 7.  Surgical data science - from concepts toward clinical translation.

Authors:  Lena Maier-Hein; Matthias Eisenmann; Duygu Sarikaya; Keno März; Toby Collins; Anand Malpani; Johannes Fallert; Hubertus Feussner; Stamatia Giannarou; Pietro Mascagni; Hirenkumar Nakawala; Adrian Park; Carla Pugh; Danail Stoyanov; Swaroop S Vedula; Kevin Cleary; Gabor Fichtinger; Germain Forestier; Bernard Gibaud; Teodor Grantcharov; Makoto Hashizume; Doreen Heckmann-Nötzel; Hannes G Kenngott; Ron Kikinis; Lars Mündermann; Nassir Navab; Sinan Onogur; Tobias Roß; Raphael Sznitman; Russell H Taylor; Minu D Tizabi; Martin Wagner; Gregory D Hager; Thomas Neumuth; Nicolas Padoy; Justin Collins; Ines Gockel; Jan Goedeke; Daniel A Hashimoto; Luc Joyeux; Kyle Lam; Daniel R Leff; Amin Madani; Hani J Marcus; Ozanan Meireles; Alexander Seitel; Dogu Teber; Frank Ückert; Beat P Müller-Stich; Pierre Jannin; Stefanie Speidel
Journal:  Med Image Anal       Date:  2021-11-18       Impact factor: 13.828

Review 8.  Crowdsourcing in Surgical Skills Acquisition: A Developing Technology in Surgical Education.

Authors:  Jessica C Dai; Thomas S Lendvay; Mathew D Sorensen
Journal:  J Grad Med Educ       Date:  2017-12

9.  Mapping of Crowdsourcing in Health: Systematic Review.

Authors:  Perrine Créquit; Ghizlène Mansouri; Mehdi Benchoufi; Alexandre Vivot; Philippe Ravaud
Journal:  J Med Internet Res       Date:  2018-05-15       Impact factor: 5.428

Review 10.  Crowdsourcing in health and medical research: a systematic review.

Authors:  Cheng Wang; Larry Han; Gabriella Stein; Suzanne Day; Cedric Bien-Gund; Allison Mathews; Jason J Ong; Pei-Zhen Zhao; Shu-Fang Wei; Jennifer Walker; Roger Chou; Amy Lee; Angela Chen; Barry Bayus; Joseph D Tucker
Journal:  Infect Dis Poverty       Date:  2020-01-20       Impact factor: 4.520

  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.