Literature DB >> 30480213

Randomized Controlled Pilot Study of Video Self-assessment for Resident Mastoidectomy Training.

Ashok R Jethwa1, Christopher J Perdoni2, Elizabeth A Kelly3, Bevan Yueh1, Samuel C Levine1, Meredith E Adams1.   

Abstract

A prospective randomized controlled pilot study was performed to determine if video self-assessment improves competency in mastoidectomy and to assess interrater agreement between expert and resident evaluations of recorded mastoidectomy. Sixteen otolaryngology residents were recorded while performing cadaveric mastoidectomy and randomized into video self-assessment and control groups. All residents performed a second recorded mastoidectomy. Performance was evaluated by blinded experts with a validated assessment scale. Video self-assessment did not lead to greater skill improvement between the first and second mastoidectomy. Interrater agreement was fair to substantial between the expert evaluators and between resident self-evaluations by recall and video review. Agreement between experts and residents was only slight to fair; residents consistently rated their performance higher than experts (P < .05). In conclusion, 1 session of video self-review did not lead to improved competence in mastoidectomy over standard practice. While experts agree on assessments, residents may overestimate their competency in performing cadaveric mastoidectomy.

Entities:  

Keywords:  mastoidectomy; neurotology; otology; resident education; self-assessment; simulation; technology; video

Year:  2018        PMID: 30480213      PMCID: PMC6239141          DOI: 10.1177/2473974X18770417

Source DB:  PubMed          Journal:  OTO Open        ISSN: 2473-974X


Mastoidectomy is a critical skill for the surgical management of the ear, temporal bone, and skull base. Historically, cadaveric temporal bone simulation has been the standard method of acquiring skills to become proficient in mastoidectomy. One novel approach to improving mastoidectomy education is through video self-assessment, similar to a postgame review by athletes. Video self-assessment of simulated and real-time procedures has proven effective in improving technical skills in several surgical specialties, including general surgery, urology, and gynecology.[1,2] In this pilot work, our goals were to (1) determine if video self-assessment improves resident skill in cadaveric mastoidectomy over standard training and (2) establish the interrater reliability between expert and resident assessments of skill based on recorded mastoidectomy.

Methods

Study Design

This was a prospective randomized pilot study among otolaryngology residents at the University of Minnesota (N = 16). Participation was voluntary. The University of Minnesota Institutional Review Board deemed this study exempt from review (No. 1501E59582). After reviewing the performance evaluation tools, all residents performed a cadaveric mastoidectomy recorded on a microscope-mounted camera, followed by self-assessment via recall. Participants were block randomized by training year to the intervention group (which received the recording) or to the control group (which did not). The intervention group reviewed the recording within 7 days and performed self-assessment. Both groups performed a second mastoidectomy 7 to 10 days after the first on a temporal bone of the same-side ear as the first session, followed by self-assessment via recall. All participants were given the video of their second mastoidectomy to review and self-assess within 7 days. Two attending neurotologists and 1 neurotology fellow served as expert assessors of each recording. They were blinded to study group, resident year, and order of mastoidectomy. Experts met prior to evaluating the videos to establish consistent evaluation techniques.

Assessment Instruments

The Task-Based Checklist (TBC) and the Global Rating Scale (GRS)—developed by Francis et al for mastoidectomy and modified to enable video review—were used for all self- and expert assessment (see Supplemental Table S1, available at www.otojournal.org/supplemental).[3] The following data were recorded: (1) time from initiation of cortical drilling to completion of mastoidectomy stages and (2) number of injuries to relevant structures. Demographic and satisfaction surveys were administered.

Statistical Analysis

We compared the change in outcome measures between the first and second mastoidectomy between the study groups with 2-tailed paired t tests. The mean of the scores assigned by the 3 expert evaluators was used for analysis. Injury counts were compared with chi-square tests. Significance was set at P < .05. Interobserver agreement was determined with weighted kappa statistics, per the criteria of Landis and Koch for interpretation of levels of agreement.[4] STATA 14.2 (StataCorp, College Station, Texas) was used for all analyses.

Results

Study groups were balanced on resident experience based on training year and mastoids previously drilled in the laboratory (mean ± SD, 6 ± 4 vs 5 ± 4) and operating room (30 ± 24 vs 31 ± 22). There were no significant differences between the first and second mastoidectomy in TBC or GRS scores, completion time of mastoidectomy stages, or injury counts between study arms or within either study arm ( ). Injury counts were low for all structures except the tegmen (11 injuries).
Table 1.

Change in Scores on Expert Evaluations and Time to Completion between First and Second Mastoidectomy in the Video Self-assessment and Control Groups.

Study Group, Mean (SD)
Video Review (Intervention)
No Video Review (Control)
ComponentFirst MastoidectomyChange[a] First MastoidectomyChange[a] P Value[b]
Task-Based Checklist
 1a. Placement of superior cut2.78 (0.93)0.04 (1.01)3.00 (0.27)0.29 (0.40).5499
 1b. Placement of canal cut2.81 (0.90)−0.07 (1.08)3.05 (0.68)0.19 (0.69).5819
 2a. Identification and definition of tegmen2.33 (0.87)−0.19 (0.53)2.43 (1.07)0.48 (1.51).2397
 2b. Sharpen posterior EAC cortex2.48 (0.85)0.19 (0.67)2.52 (1.02)0.24 (0.9).8941
 2c. Define sigmoid sinus and sinodural angle2.48 (1.02)0.07 (1.10)2.81 (0.72)−0.05 (1.08).8283
 3a. Deepen dissection at sinodural angle2.48 (0.77)0.07 (0.64)2.43 (1.01)−0.24 (1.08).4827
 3b. Open antrum from posterior to anterior2.37 (0.73)−0.22 (0.60)2.29 (0.97)−0.14 (1.03).8496
 3c. Atraumatic exposure of short process of incus2.48 (0.80)0.22 (0.87)2.48 (0.72)0.62 (1.03).4153
 4a. View posterior EAC en face2.22 (0.83)−0.07 (1.02)2.14 (0.79)0.00 (0.77).8759
 4b. Use side/front of appropriate bur2.67 (0.76)0.19 (0.67)2.29 (0.95)−0.24 (1.15).3701
 4c. Saucerization2.37 (0.84)0.04 (0.65)2.10 (0.69)−0.29 (1.25).5149
Global Rating Scale
 1. Use of otologic drills2.59 (0.74)0.07 (1.01)2.33 (1.15)−0.33(1.33).4973
 2. Use of irrigation2.85 (0.5)0.22 (0.76)3.0 (0.51)0.10 (0.37).6935
 3. Use of microscope2.70 (0.65)0.00 (0.91)2.76 (1.10)−0.05 (1.35).9341
 4. Respect for surgical limits2.41 (1.23)−0.04 (1.12)2.57 (1.15)0.10 (1.63).8501
 5. Time and motion2.63 (0.98)0.15 (0.63)2.52 (1.14)−0.24 (1.33).4520
 6. Flow of operation2.74 (1.01)0.22 (0.65)2.71 (1.11)−0.05(1.42).6180
 7. Overall surgical performance2.44 (1.03)0.0 (0.69)2.48 (1.14)0.05 (1.24).9232
Time from cortex to structure, min
 Tegmen12.7 (10.5)−1.0 (3.5)11.5 (6.4)−2.2 (6.8).7074
 Sigmoid14.2 (8.4)−4.4 (10.6)16.5 (10.0)−7.1 (11.6).7154
 Incus19.2 (8.2)−0.3 (4.9)21.4 (9.0)−3.8 (5.5).3107
 Total26.6 (13.5)−3.1 (7.3)37.7 (22.5)−16.5 (22.9).1792

Abbreviation: EAC, external auditory canal.

Change between first and second mastoid. A negative value indicates that the score in the second mastoidectomy was lower (worse).

P values are for the paired t tests comparing the change from the first to second mastoidectomy between the study groups.

Change in Scores on Expert Evaluations and Time to Completion between First and Second Mastoidectomy in the Video Self-assessment and Control Groups. Abbreviation: EAC, external auditory canal. Change between first and second mastoid. A negative value indicates that the score in the second mastoidectomy was lower (worse). P values are for the paired t tests comparing the change from the first to second mastoidectomy between the study groups. Interrater agreement ( ) was fair to substantial among the expert evaluators (κ = 0.23-0.62) with highest agreement on overall surgical performance. Interrater agreement was fair to substantial among resident self-evaluations by recall and video review (κ = 0.40-0.78). There was only slight to fair agreement between expert ratings and resident ratings (κ = 0.03-0.25). For all TBC and GRS items, resident ratings were 0.41 to 1.51 points higher than the mean expert ratings (all P < .007; ).
Table 2.

Weighted Kappa Statistics for Interrater Agreement.

Weighted Kappa for Rater Agreement
ComponentResidents[a] Experts[b] Experts vs Residents[c]
Task-Based Checklist
 1a. Placement of superior cut0.640.310.12 (0.05 to 0.18)
 1b. Placement of canal cut0.420.440.17 (0.15 to 0.18)
 2a. Identification and definition of tegmen0.510.510.21 (0.17 to 0.27)
 2b. Sharpen posterior EAC cortex0.560.360.13 (0.09 to 0.15)
 2c. Define sigmoid sinus and sinodural angle0.440.460.09 (0.02 to 0.18)
 3a. Deepen dissection at sinodural angle0.570.450.18 (0.10 to 0.29)
 3b. Open antrum from posterior to anterior0.400.260.12 (0.04 to 0.17)
 3c. Atraumatic exposure of short process of incus0.770.480.13 (0.09 to 0.18)
 4a. View posterior EAC en face0.780.400.12 (0.06 to 0.17)
 4b. Use side/front of appropriate bur0.570.380.08 (0.07 to 0.09)
 4c. Saucerization0.630.390.03 (0.02 to 0.04)
Global Rating Scale
 1. Use of otologic drills0.520.500.03 (0.01 to 0.06)
 2. Use of irrigation0.560.230.04 (–0.02 to 0.15)
 3. Use of microscope0.730.450.07 (0.02 to 0.1)
 4. Respect for surgical limits0.510.610.22 (0.21 to 0.24)
 5. Time and motion0.450.420.18 (0.12 to 0.27)
 6. Flow of operation0.640.460.25 (0.16 to 0.32)
 7. Overall surgical performance0.680.620.19 (0.13 to 0.26)

Abbreviation: EAC, external auditory canal.

Self-recall vs self-video assessment.

Agreement among 3 expert assessments.

Weighted kappa was calculated for agreement between each expert (n = 3) and resident self-video assessments. The mean (range) of the 3 kappas is presented here.

Figure 1.

Expert vs resident self-assessment of recorded mastoidectomy performance with the Task-Based Checklist and Global Rating Scale. EAC, external auditory canal. Mean values are presented, with error bars indicating SD.

Weighted Kappa Statistics for Interrater Agreement. Abbreviation: EAC, external auditory canal. Self-recall vs self-video assessment. Agreement among 3 expert assessments. Weighted kappa was calculated for agreement between each expert (n = 3) and resident self-video assessments. The mean (range) of the 3 kappas is presented here. Expert vs resident self-assessment of recorded mastoidectomy performance with the Task-Based Checklist and Global Rating Scale. EAC, external auditory canal. Mean values are presented, with error bars indicating SD. Residents rated satisfaction with video assessment highly (4 ± 1.4 out of 5), and 78% said that they would repeat the study.

Discussion

In this study, 1 episode of video self-review did not produce improved competence in mastoidectomy over standard training. Expert and resident assessments conflicted, as residents consistently rated themselves higher than experts did. Our findings should be considered in the context of other literature. Malik et al demonstrated a negative association between time spent in the temporal bone laboratory and resident competence in mastoidectomy.[5] In their study and ours, residents did not receive expert feedback or coaching. Expert feedback is critical for complex task learning. Hu et al demonstrated that, in contrast to standard practices, video-based coaching sessions of surgical procedures generate more questions and detailed discussions between attendings and residents about intraoperative decision making.[6] Feedback is also important for development of accurate self-estimations of skill.[7-9] Thus, these findings suggest that mastoidectomy simulation without feedback may not promote new skill development or it may lead to bad habits, as residents may not recognize or correct their own technical errors and inefficiencies. Video review also allows experts to reflect on educational techniques, identifying areas that require focus. For example, we increased our emphasis on early definition of tegmen contours to address the high rates of injury observed. Furthermore, we learned that recordings provide an accessible and efficient means of observing resident drilling and may augment evaluation and feedback when attendings cannot be present during each laboratory dissection. The primary limitation of our study is its small sample size, drawn from a single institution. Enrollment in multiple programs would increase study power and external validity. Furthermore, as self-video review was limited to 1 session, additional video review sessions with multiple cadaveric temporal bone dissections may lead to improvements that could not be detected in this study. Future work will determine if joint video review with an expert may afford greater educational value.

Author Contributions

Ashok R. Jethwa, substantial contributions to the conception and design of the work, drafting the work and revising it critically for important intellectual content, final approval of the version to be published, agreement to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved; Christopher J. Perdoni, substantial contributions to the acquisition of the data for the work; revising it critically for important intellectual content, final approval of the version to be published, agreement to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved; Elizabeth A. Kelly, substantial contributions to the acquisition and interpretation of the data for the work; revising it critically for important intellectual content; final approval of the version to be published; agreement to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved; Bevan Yueh, substantial contributions to the design of the work; revising the work critically for important intellectual content; revising it critically for important intellectual content; final approval of the version to be published; agreement to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved; Samuel C. Levine, substantial contributions to the acquisition of the data for the work; revising it critically for important intellectual content; final approval of the version to be published, agreement to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved; Meredith E. Adams, substantial contributions to the conception and design of the work, drafting the work and revising it critically for important intellectual content, final approval of the version to be published, agreement to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved.

Disclosures

Competing interests: Samuel C. Levine, Grace Medical—receives royalties for Levine disposable suction irrigation tubing. Sponsorships: None. Funding source: American Academy of Otolaryngology—Head and Neck Surgery Resident Research Grant. The funding source had no role in study design and conduct; collection, analysis, and interpretation of the data; or writing or approval of the manuscript. Click here for additional data file. Supplemental material, OPN770417_Supplemental_Material_CLN for Randomized Controlled Pilot Study of Video Self-assessment for Resident Mastoidectomy Training by Ashok R. Jethwa, Christopher J. Perdoni, Elizabeth A. Kelly, Bevan Yueh, Samuel C. Levine and Meredith E. Adams in OTO Open: The Official Open Access Journal of the American Academy of Otolaryngology-Head and Neck Surgery Foundation
  9 in total

1.  The influence of self-deception and impression management upon self-assessment in oral surgery.

Authors:  A W Evans; R M A Leeson; T R O Newton John; A Petrie
Journal:  Br Dent J       Date:  2005-06-25       Impact factor: 1.626

2.  Determinants of resident competence in mastoidectomy: role of interest and deliberate practice.

Authors:  Mohammad U Malik; David A Diaz Voss Varela; EunMi Park; Hamid Masood; Kulsoom Laeeq; Nasir I Bhatti; Howard W Francis
Journal:  Laryngoscope       Date:  2013-07-22       Impact factor: 3.325

3.  "I'll never play professional football" and other fallacies of self-assessment.

Authors:  Kevin W Eva; Glenn Regehr
Journal:  J Contin Educ Health Prof       Date:  2008       Impact factor: 1.355

4.  Pilot testing of an assessment tool for competency in mastoidectomy.

Authors:  Kulsoom Laeeq; Nasir I Bhatti; John P Carey; Charles C Della Santina; Charles J Limb; John K Niparko; Lloyd B Minor; Howard W Francis
Journal:  Laryngoscope       Date:  2009-12       Impact factor: 3.325

5.  Video self-assessment augments development of videoscopic suturing skill.

Authors:  Ramin Jamshidi; Teresa LaMasters; Dan Eisenberg; Quan-Yang Duh; Myriam Curet
Journal:  J Am Coll Surg       Date:  2009-09-19       Impact factor: 6.113

6.  An application of hierarchical kappa-type statistics in the assessment of majority agreement among multiple observers.

Authors:  J R Landis; G G Koch
Journal:  Biometrics       Date:  1977-06       Impact factor: 2.571

7.  Video-based peer feedback through social networking for robotic surgery simulation: a multicenter randomized controlled trial.

Authors:  Stacey C Carter; Alexander Chiang; Galaxy Shah; Lorna Kwan; Jeffrey S Montgomery; Amer Karam; Christopher Tarnay; Khurshid A Guru; Jim C Hu
Journal:  Ann Surg       Date:  2015-05       Impact factor: 12.969

8.  Complementing Operating Room Teaching With Video-Based Coaching.

Authors:  Yue-Yung Hu; Laura M Mazer; Steven J Yule; Alexander F Arriaga; Caprice C Greenberg; Stuart R Lipsitz; Atul A Gawande; Douglas S Smink
Journal:  JAMA Surg       Date:  2017-04-01       Impact factor: 14.766

Review 9.  Self versus external assessment for technical tasks in surgery: a narrative review.

Authors:  Boris Zevin
Journal:  J Grad Med Educ       Date:  2012-12
  9 in total
  2 in total

1.  The effect of structured self-assessment in virtual reality simulation training of mastoidectomy.

Authors:  Steven Arild Wuyts Andersen; Mads Guldager; Peter Trier Mikkelsen; Mads Sølvsten Sørensen
Journal:  Eur Arch Otorhinolaryngol       Date:  2019-09-20       Impact factor: 2.503

2.  Understanding the effects of structured self-assessment in directed, self-regulated simulation-based training of mastoidectomy: A mixed methods study.

Authors:  Steven Arild Wuyts Andersen; Martin Frendø; Mads Guldager; Mads Sølvsten Sørensen
Journal:  J Otol       Date:  2019-12-23
  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.