| Literature DB >> 34873653 |
Tora Rydtun Haug1,2, Mai-Britt Worm Ørntoft1,3, Danilo Miskovic4, Lene Hjerrild Iversen3, Søren Paaske Johnsen5, Anders Husted Madsen6.
Abstract
BACKGROUND: In laparoscopic colorectal surgery, higher technical skills have been associated with improved patient outcome. With the growing interest in laparoscopic techniques, pressure on surgeons and certifying bodies is mounting to ensure that operative procedures are performed safely and efficiently. The aim of the present review was to comprehensively identify tools for skill assessment in laparoscopic colon surgery and to assess their validity as reported in the literature.Entities:
Keywords: Assessment tool; Colon surgery; Competency; Laparoscopy; Surgical education; Technical skills
Mesh:
Year: 2021 PMID: 34873653 PMCID: PMC8847271 DOI: 10.1007/s00464-021-08914-z
Source DB: PubMed Journal: Surg Endosc ISSN: 0930-2794 Impact factor: 4.584
Definitions of validity sources.
Adopted from Beckman et al. [16] and Ghaderi et al. [13] Modified for the scope of this review
| Domain | Definition | Score | Description | Examples |
|---|---|---|---|---|
| The extent to which the tool’s content relates to the construct it intends to measure | 0 | No data regarding the developing process | ||
| 1 | Expert judgment with limited data regarding the tool content | Expert judgment | ||
| 2 | Listing assessment items for the tool content with some references to a panel of experts, limited description of the developing process | Structured task analysis, hierarchical task analysis | ||
| References to a previously validated tool | Based on previously validated tools | |||
| 3 | Well-defined developing process, both theoretical basis for the chosen items and systematic review by experts | Delphi-method, pilot study | ||
| The analysis of the responses given by the individual assessors and interpretation of the reported results | 0 | No data regarding the response process | ||
| 1 | Limited data reported. Use of an assessment tool without discussing the impact of the differences in response processes | User manuals | ||
| 2 | Some data regarding different responses of assessors. Some data about systems that reduce variation between respondents | Structured assessor training before the assessment process | ||
| 3 | Multiple sources of data examining response error through critical examination of response processes and respondents | Validation of initial scores (pilot study), evaluation of response error after structured assessor training | ||
| The extent to which individual items describe the underlying constructs, often reported by measures of inter-rater reliability, internal consistency and generalizability | 0 | No data regarding internal structure | ||
| 1 | Limited data regarding internal structure, references to a single inter-rater reliability measure | Simple measures of inter-rater reliability (ICC or Cronbach alpha) | ||
| 2 | A few measures of reliability reported, insufficiently item analysis | Inter-rater reliability coefficient combined with a single measure of interitem or intertest reliability | ||
| 3 | Multiple measures of reliability including inter-rater reliability and item-analysis (interitem reliability, inter-test reliability, item response theory) | Generalizability theory analysis, item response theory | ||
| Correlation between the assessment scores and other outcomes or scoring systems relevant to the construct being measured | 0 | No data regarding relations to other variables | ||
| 1 | Correlation of scores to outcomes with limited theoretical importance, references to a single measure of validity | Compare level of training to score achieved with the assessment tool | ||
| 2 | Correlation of scores to outcomes with theoretical importance, references to a few measures of validity | Correlation with level of training and clinical data (operative time, patient outcomes etc.) | ||
| 3 | Correlation or no correlation between important theoretical outcomes or scores of the same construct | Correlation with training level, clinical data and other performance assessment tools, generalizability evidence | ||
The impact of the assessment tool and future use | 0 | No data or discussion regarding consequences | ||
| 1 | Limited data, merely a discussion about future use | Describing feasibility and potential future use (data on assessment time, post assessment survey) | ||
| 2 | Some descriptions of consequences of assessment for learners, often supported by incomplete data | Describing educational impact (formative / summative feedback, learning curve of trainees) | ||
| 3 | Clear description of consequences of assessments and the impact on interpretation of scores and intended future use, supported by data | Criterion-referenced score (pass/fail-scores), cut-of scores for licensing purposes, predictive models |
Fig. 1Flowchart of the included studies. AT: assessment tool, lap. colon: laparoscopic colon, other: language, review, protocol paper, editorial, conference abstract, commentary
Characteristics*
| Assessment tools | 14 (100%) | |
|---|---|---|
| Global rating scale | 5 (36) | |
| Error-based rating scale | 1 (7) | |
| Procedure-specific tool | 8 (57) | |
| 2005–2010 | 4 (29) | |
| 2011–2015 | 6 (43) | |
| 2016–2020 | 4 (29) | |
| United Kingdom | 5 (36) | |
| Canada | 4 (29) | |
| United States | 3 (21) | |
| Japan | 2 (14) | |
| Modified | 5 (36) | |
| Original | 7 (50) | |
| Modified and original | 2 (14) | |
| Video | 7 (50) | |
| Direct observation | 5 (36) | |
| Both video and direct observation | 1 (7) | |
| Unspecified | 1 (7) | |
*A total of 22 studies were identified which included 14 different assessment tools. Only the paper describing the developing process has been included for the tools described in multiple papers
Descriptive data of assessment tools
| Tool | Abb | Year | Items | Versions | Score range | Original or modified | ||
|---|---|---|---|---|---|---|---|---|
| Global Rating Scale | ||||||||
| 1 | Sidhu et al. [ | SAS | 2006 | 11 | 1 | 11–55 | Previously validated GOALS and OSATS | |
| 2 | Wohaibi et al. [ | OpRate | 2007 | 6 | 1 | 1–4 (m) | Original | |
| 3 | Niitsu et al. [ | OSATS | 2012 | 7 | 1 | 7–35 | Previously validated OSATS | |
| 4 | Jenkins et al. [ | GMAS | 2016 | 12 | 1 | 0–45 | Original GMAS and previously validated DOPS | |
| 5 | Watanabe et al. [ | IRT-GOALS | 2017 | 5 | 1 | 5–25 | Previously validated GOALS | |
| Error-based rating scale | ||||||||
| 6 | Miskovic et al. [ | OCHRA | 2012 | 3 | 1 | 3–25 | Previously validated GAS, and OCHRA | |
| Procedure-specific tool | ||||||||
| 7 | Dath. et al. [ | OCRS | 2003 | 7 | 1 | 1–5(m) | Original OCRS and previously validated OSATS | |
| 8 | Sarker et al. [ | TSALC | 2010–2011 | 7–9 | 3% | 35–45 | Original | |
| 9 | Palter et al. [ | PSET | 2011–2012 | 18 or 18/19 | 2 & | 50–80 or 50–90 | Original | |
| 10 | Miskovic et al. [ | GAS | 2011 | 12 | 1 | 1–72 | Original | |
| 11 | Miskovic et al. [ | CAT | 2013 | 16 | 1 | 1–4 (m) | Original | |
| 12 | Glarner et al. [ | CT | 2013 | 8 | 1 | 1–5 (m) | Original CT and modified OSATS and NOTSS | |
| 13 | Champagne et al. [ | ASCRS | 2017 | 9 | 1 | 1–5 (m) | Previously validated OSATS and OCHRA | |
| 14 | Nakayama et al. [ ASLAC score | ASLAC | 2020 | 35 | 1 | NR | Original | |
Abb: Abbreviation Year: year of publication Items: number of statements evaluated on a rating scale. If the tool includes general, non-technical and procedure-specific items, only the procedure-specific items will be considered Versions: number of versions available for laparoscopic colon procedures NR: not reported (m) mean
*Only number from the first validation paper has been reported
Data describing the validation process of assessment tools
| Tool | Partici-pants | Cases† | Procedure evaluated | Validation setting | Live / video | Assessor | Assessors in total | |
|---|---|---|---|---|---|---|---|---|
| Global Rating Scale | ||||||||
| 1 | Sidhu et al. [ | 22 (T) | 22 | Lap sigmoid colectomy | Laboratory (animal) | video | 2 | 2 |
| 2 | Wohaibi et al. [ | 29 (T) | 579† | Lap colon resections † | Workplace | live | 1 | 33 |
| 3 | Niitsu et al. [ | 10 (T) | 757† | Lap colon resections † | Workplace | live | 1 | 12 |
| 4 | Jenkins et al. [ | 8 (C) | 750 | Lap colon and rectum resections | Workplace | video, live | 1 | 2 |
| 5 | Watanabe et. al [ | 396 (T + C) | 396† | Lap colon and rectum resections † | Workplace | live | 1 | NR |
| Error-based rating scale | ||||||||
| 6 | Miskovic et al. [ | 21 (C) | 33 | Lap colon resections (right and left colectomies) | Workplace | video | 1–3 (2 for sum-mative feedback) | 16 |
| Procedure-specific tool | ||||||||
| 7 | Dath. et al. [ | 29 (T) | 58† | Lap low anterior resections | Laboratory (animal) | video | 2 | 10 |
| 8 | Sarker et al. [ | 14 (T + C) | 84 | Lap colon and rectum resections (right hemicolectomies, sigmoid resections and anterior resections) | Workplace | video | 2 | 2 |
| 9 | Palter et al. [ | 37 (T + C) | 37 | Lap right colectomies and lap sigmoid colectomies | Workplace | video | 2 | 2 |
| 10 | Miskovic et al. [ | 52 (C) | 333 | Lap colon and rectum resections (right hemicolectomies, sigmoid resections, anterior resections, low anterior resection, total and subtotal colectomies and assisted abdominoperineal resections) | Workplace | live | 1–2 (self-evaluation by subject) | 30 |
| 11 | Miskovic et al. [ | 31 (C) | 54 | Lap right and left colectomies | Workplace | video | 2–3 | 27 |
| 12 | Glarner et al. [ | 16 (T) | 63 | Lap segmental colon resections | Workplace | live | 1 | 4 |
| 13 | Champagne et al. [ | 24 (T + C) | 24 | Lap right hemicolectomies | Workplace, | video | 5 | 20 |
| 14 | Nakayama et al. [ ASLAC score | NR | NR | NR | NR | NR | NR | NR |
Participants: number of participating trainees (T) and consultants (C) Trainees: doctors purchasing a career in surgery who have not completed their residency as post-graduate year surgeons Consultants: surgeons who have completed their post-graduate residency and those who are specialised in operating on colon and rectum Cases: number of live or video recorded operations included in the analysis Lap: laparoscopic Assessors pr. case: number of assessor on each procedure Assessors in total: assessors contributing to the scoring process NR: not reported
†Other surgical procedures also reported
*Only number from the first validation paper has been reported
Evidence of validity
| Tool | Content | Response process | Internal structure | Relations to other variables | Conse-quences | Total | |
|---|---|---|---|---|---|---|---|
| Global Rating Scale | |||||||
| 1 | Sidhu et al. [ | 2 | 1 | 2 | 1 | 1 | |
| 2 | Wohaibi et al. [ | 0 | 1 | 1 | 1 | 0 | |
| 3 | Niitsu et al. [ | 2 | 2 | 0 | 1 | 1 | |
| 4 | Jenkins et al. [ | 2 | 2 | 0 | 2 | 3 | |
| 5 | Watanabe et al.[ | 2 | 0 | 3 | 1 | 3 | |
| Error-based rating scale | |||||||
| 6 | Miskovic et al. [ | 2 | 2 | 2 | 3 | 3 | |
| Procedure-specific tool | |||||||
| 7 | Dath. et al. [ OCRS | 2 | 1 | 2 | 1 | 1 | |
| 8 | Sarker et al. [ | 2 | 0 | 1 | 1 | 1 | |
| 9 | Palter et al. [ | 3 | 1 | 1 | 1 | 1 | |
| 10 | Miskovic et al. [ | 2 | 1 | 2 | 1 | 3 | |
| 11 | Miskovic et al. [ | 3 | 1 | 3 | 3 | 3 | |
| 12 | Glarner et al. [ | 3 | 1 | 0 | 1 | 1 | |
| 13 | Champagne et al. [ | 3 | 2 | 2 | 2 | 3 | |
| 14 | Nakayama et al. [ ASLAC score | 3 | 0 | 0 | 0 | 0 | |
*Scoring system: 0: the study provided no discussion or data, 1: the study provided limited data that support validity evidence, 2: the study provided some data (intermediate) that support validity evidence, 3: the study provided multiple data that support validity evidence
*The highest level of validity for the respective studies is reported