Literature DB >> 35515747

Low versus high level of physical resemblance in simulation for the acquisition of basic surgical skill: a meta-analysis.

Fabrizio Consorti1, Gianmarco Panzera1.   

Abstract

Background: Many studies explored the use of simulation in basic surgical education, with a variety of devices, contexts and outcomes, with sometimes contradictory results.
Objectives: The objectives of this meta-analysis were to focus the effect that the level of physical resemblance in a simulation has on the development of basic surgical skill in undergraduate medical students and to provide a foundation for the design and implementation of a simulation, with respect to its effectiveness and alignment with the learning outcomes. Study selection: We searched PubMed and Scopus database for comparative randomised studies between simulations with a different level of resemblance. The result was synthesised as the standardised mean difference, under a random effect model. Findings: We selected 12 out of 2091 retrieved studies, reporting on 373 undergraduate students (mean of subjects 15.54±6.89). The outcomes were the performance of simple skills and the time to complete a task. Two studies reported a scoring system; seven studies reported time for a task; and three studies reported both. The total number of measures included in the meta-analysis was 456 for score and 504 for time. The pooled effect size did not show any significant advantage in a simulation of a high level of physical resemblance over a lower level, both for the scoring system (-0.19, 95% CI -0.44 to 0.06) and for time (-0.14, 95% CI -0.54 to 0.27).
Conclusion: Simulations with a low level of physical resemblance showed the same effect as the simulation using a higher level of resemblance on the development of basic surgical skills in undergraduate students. © Author(s) (or their employer(s)) 2021. Re-use permitted under CC BY-NC. No commercial re-use. See rights and permissions. Published by BMJ.

Entities:  

Keywords:  medical education research; simulation; surgery general; surgical education; undergraduate education

Year:  2021        PMID: 35515747      PMCID: PMC8936611          DOI: 10.1136/bmjstel-2020-000797

Source DB:  PubMed          Journal:  BMJ Simul Technol Enhanc Learn        ISSN: 2056-6697


Introduction

‘The earlier surgical skills training starts the better’. This was one of the 12 tips Kneebone1 listed more than 20 years ago, highlighting the importance of an early beginning of surgical training in a medical curriculum. Actually, many international curricula2 and core competence inventories3 include basic surgical skill as a component of professional competence. Clerkship remains the cornerstone of surgical undergraduate education,4 5 conjugated with cognitive methods like case-based learning.6 The development of more technical surgical skills is also based on methods like the boot camps,7 special programmes8 or simulation.9 10 Simulation is a highly effective training method, and many factors influence its effectiveness.11 Among the others, many authors focused on the level of fidelity in various clinical domains.12–14 Hamstra et al 15 questioned the term ‘fidelity’ and argued that the concept of fidelity is multidimensional. They suggested using the terms physical resemblance, to denote the physical structure of the device used in simulation, and functional task alignment, to denote the coherence of the whole context of simulation with the expected learning outcomes. Finally, they recommended giving more emphasis to the transfer of learning, that is, the ability to apply the skill trained in a context to a different context. The use of simulation in surgical education has been extensively studied, mainly in postgraduate training for mini-invasive surgery. In this regard, five metanalyses have been published,16–20 but none of them explored the effect of simulation in the development of surgical skill in undergraduate students, despite many studies exploring the use of simulation in this field, with a variety of devices, context and outcomes. There are theoretical reasons—discussed in the next section—to hypothesise that the leaning outcomes included in the class of basic surgical skill can be achieved with a low level of physical resemblance. The objective of this meta-analysis was to review and synthetise the available evidence about the effect that the level of physical resemblance in a simulation has on the development of surgical skill in undergraduate medical students, with the goal to provide a foundation for the design and implementation of simulation, both with respect to the optimal use of resources and to the alignment to the task. The design of this meta-analysis has been registered in the PROSPERO database (CRD42019128061).

Theoretical background

In this section, first, we give a definition of basic surgical skill and of undergraduate student, then we consider some of the educational theories that can predict the role of the characteristics of simulation in its effectiveness and we briefly discuss them. We considered as basic surgical skills both simple tasks like gowning and gloving, suturing, knot tying2 21 and a set of psychomotor abilities like dexterity and coordination of the two hands, or the ability to work with a 2-D view. This last set of skills is important because of the increasing importance that minimally invasive surgery has.22 We considered as undergraduate students those attending a medical school according to the general model described by Wijnen-Meijer et al. 23 Behaviourism is one of the most frequent theoretical positions to frame simulation. It posits that the outcome of learning is the modification of behaviour, regardless of the implicated cognitive mechanisms. Teaching is then providing an adequate stimulus to obtain the desired response.24 An effective training starts from easy tasks and progresses through an intentional sequence up to proficiency. Deliberate practice and mastery learning are typical behaviourist instructional methods. Also, the use of a part-task trainer in simulation is justified in a behaviourist approach. Cognitivism is a class of theories that explore the mental mechanisms in learning. In this class, situated cognition theory maintains that learning emerges from the interplay of the individual cognitive process and the social, cultural and physical environment in which the process takes place.25 This theory justifies the effectiveness of simulations with a high level of physical resemblance, in which the action takes place in an environment that recalls the real professional situation. Another relevant contribution of the cognitivist perspective is the concept of mental workload. Our mind has a limited capacity of working memory, and an overload of cognitive and sensory information can alter mental performances like learning or taking clinical decisions. Naismith and Cavalcanti26 reviewed the effect of cognitive load in simulation, showing that an excess of load impairs learning. The learning model known as the challenge point framework27 builds on both behaviourist and a cognitivist perspectives. This model describes the relationship between the performance and the difficulty of a task, defined as the physical or cognitive challenge posed by a motor problem. The model proposes a progression from simpler to more difficult tasks as the most effective strategy for learning. Based on these premises, if basic surgical skills are the building blocks of a specific professional competence to be further developed in the future of graduate education, then we hypothesise that a simulation aligned to these basic outcomes and based on a low level of physical resemblance is as effective as simulation with a higher level of resemblance and that reproduces a realistic professional context.

Materials and methods

We planned, conducted and reported this study according to Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) standards of quality.28

Study eligibility

This meta-analysis considered trials in which undergraduate students were randomised to use different equipment for the simulation of surgical skills. For laparoscopic simulation, we considered a simulation in which either immersive virtual reality or 3-D vision was used as having a higher level of physical resemblance than the simulations in which a box trainer was used, either with direct vision or based on a conventional video system. For non-laparoscopic simulation, a high physical resemblance was defined as the use of a biological model as opposed to the use of a simplified synthetic model. We excluded studies about endoscopic skill (gastroscopy and bronchoscopy) because these skills did not fit with the definition we gave of basic surgical skills. Studies comparing simulation with a no-intervention or no-simulation group were excluded. In studies with three arms, we considered only the two arms comparing devices that we interpreted as high or low resemblance. The articles had to provide the figures needed to compute the overall effect size, and the considered outcome measures were both any scoring system for skill and time for a task. We selected articles in any language, without time limits.

Sources and search strategy

We searched PubMed, Scopus and ERIC electronic databases. For PubMed, the search strategy used different combinations of Medical Subject Headings (MeSH) terms like (simulation training(MeSH Terms)), (surgery(MeSH Terms)), (education, medical, undergraduate(MeSH Terms)) with free text strings like “suturing”, “knot”, “basic skill”, “fidelity”. For Scopus and ERIC, we adopted a similar strategy based on the same keywords. We also used the references of the retrieved reviews to look for other possible relevant articles. The search was updated until June 2020. The result of the queries was uploaded into the Zotero reference management system.

Study selection

After removal of duplicates, the two authors (FC and GP) independently checked each title and abstract to identify eligible studies. We retrieved then a full-text copy of each selected article for a final decision on the fulfilment of the eligibility criteria. Any disagreements on the selection were resolved by discussion among the authors.

Data collection process

We analysed each selected article to extract the relevant information, according to the Patient Intervention Comparison Outcome model and the information was recorded in a form. In this meta-analysis, the subjects were undergraduate students; as already defined, the intervention was the description of the used device, the comparison was the level of resemblance, and the outcome was the set of trained skills, with their measure. When a study reported more than one outcome (eg, both score and time), the outcomes were separately recorded. When a study reported more than one measure for the same outcome, all the values were used to compute the overall effect size if they pertained to different tasks. If they were elements of a global score (like in GOALS score), only the total score was recorded. Finally, data were entered into the Review Manager software (RevMan V.5) by Cochrane collaboration.

Risk of bias

The quality of the evidence provided by the selected articles was assessed according to the GRADE guidelines,29 embedded in the RevMan software. The authors independently assessed the articles using the GRADE rules (study limitations, imprecision, inconsistency, indirectness and publication bias) and computing the score according to the four GRADE quality levels, from high to very low. Any disagreement on the evaluation of quality of an article of one grade or more were resolved by discussion among the authors. The publication bias was assessed with the funnel plot, observing if the distribution of the outcome measures was symmetrical around the midline.

Synthesis of results

We classified the studies according the two outcomes of scoring system and time for a task. The synthesis was computed as the standardised mean difference of effect. Variability was expressed as 95% CI. We adopted the random effect model because we expected a moderate to high heterogeneity among studies. Heterogeneity was expressed with the I2 statistic, which describes the percentage of variation across studies. It has been suggested that a value below 40% represents low inconsistency; between 40% and 60% represents moderate inconsistency; and over 60% represents high inconsistency. We used the RevMan V.5 software for all statistical calculations and for the generation of the forest plots.

Results

Flow of search

Our initial search found 2091 articles. One hundred and one met the inclusion criteria after the first screening, but after reading the abstract and the full text, only 12 were eligible for the meta-analysis. They were all written in English. Figure 1 shows the flow of search and the reasons to discard the articles not fulfilling the inclusion criteria. Nine studies reported time as the outcome; three reported a scoring system; and two reported both outcomes.
Figure 1

Flowchart of the search of articles.

Flowchart of the search of articles.

Study characteristics

Overall, the selected 12 studies were randomised trials, reporting on 373 undergraduate students (mean of subjects 15.54±6.89). All the students were defined as ‘novice students’, without any previous experience of surgery. The four studies reported a scoring system as the outcome, measured from one to six different items; the 10 studies reported time for a task measured the time for one to four different tasks. Hence, the total number of measures included in the meta-analysis was 456 for score and 504 for time. Five studies trained simple laparoscopic psychomotor skill, like pick and place jacks or thread the rings; five studies trained basic laparoscopic tasks (suturing and knot tying or cutting); one study trained part of a laparoscopic cholecystectomy; and one study trained simple interrupted and subdermal interrupted sutures. Table 1 summarises the characteristics of the included studies.
Table 1

Characteristics of the selected studies

ArticleSubjectsSkillComparisonOutcomeMeasureResult
Brinkman et al 30 18 vs 18Laparoscopic cholecystectomy in a pig modelBT versusVR trainerPerformance scoreGOALSfive custom itemsBT better than VR
Denadai et al 31 12 vs 12Simple interrupted sutures and subdermal interrupted suturesSynthetic versus animal bench modelPerformance scoreEfficiencyGlobal Rating ScaleMean number of stitches in given timeNo significant difference
Hassan et al 36 15 vs 15Laparoscopic psychomotor skillBT versus VRTimeSecondsVR better than BT
Kanumuri et al 37 8 vs 8Laparoscopic suturing and knot tyingComputer enhanced versus VRTimeSecondsNo significant difference
Kotari et al 200238 11 vs 13Laparoscopic psychomotor skillPelvic trainer versus VRTimeSecondsNo significant difference
Lehmann et al 39 12 vs 12Laparoscopic psychomotor skillConventional video trainer versus VRTimeSecondsNo significant difference
Lin et al 40 40 vs 20Laparoscopic suturing and knot tyingDirect vision training versus video-based trainingTimeSecondsVideo-based better than BT
Madan and Frantzides34 14 vs 17Laparoscopic psychomotor skillConventional video trainer versus VRPerformance scoreTimeCustom metricsSecondsNo significant difference
Nemani et al 32 6 vs 7Laparoscopic pattern cutting taskFundamentals of Laparoscopic Surgery (BT) versusVirtual Basic Laparoscopic Skills Trainer (VR)Performance scoreTimeCustom metricsSecondsNo significant difference
Tanoue et al 33 20 vs 20Laparoscopic psychomotor skillSuturing and knot tyingBT versusProcedicus MIST (VR)Efficiency*TimeErrorsMovementsSecondsBT better than VR
Van Bruwaene et al 35 21 vs 21Laparoscopic suturingBT versus VRTimeSecondsBT better than VR
Youngblood et al 41 17 vs 16Three laparoscopic tasks in a pig modelTower Trainer (BT) versusLapSim (VR)Performance scoreTimeCustom global scoreSecondsVR better than BT

BT, box trainer; VR, virtual reality.

Characteristics of the selected studies BT, box trainer; VR, virtual reality.

Study quality

Overall, the quality of the selected studies was from moderate to high: participants were randomised; we could not detect selection bias; some studies reported two different outcomes; other studies reported incomplete data for one of the outcomes, and this outcome was not considered in the meta-analysis. These studies were anyway considered for the outcome with complete data. A frequent bias was that the blinding of outcome assessment was not reported. The sample size was generally small; only two studies enrolled more than 40 students. The distribution of outcomes in the funnel plot was symmetric, both for scoring system and for time, so we could infer the absence of a publication bias.

Comparison of scoring systems

Four studies measured the skill as the outcome with a scoring system. Two of these studies30 31 used a validated score (GOALS and the Global Rating Scale); two32 33 used an original scoring system. One more study used a scoring system, but the article did not report the detailed figures,34 so it was not possible to use it for the calculation of the pooled effect size. One study reported a significant larger effect size for high resemblance35; one study reported the superiority of low resemblance30; the other two31 32 could not show any significant difference. The pooled effect size (figure 2) was small and did not show any significant advantage of a simulation with a high level over a low level of resemblance (−0.19, 95% CI −0.44 to 0.06). Heterogeneity was moderate (I2=42%).
Figure 2

Forest plot of the comparison between simulations with a high versus low level of physical resemblance for the outcome score. The pooled effect size is represented by the black diamond. CI, confidence interval; IV, interval variable.

Forest plot of the comparison between simulations with a high versus low level of physical resemblance for the outcome score. The pooled effect size is represented by the black diamond. CI, confidence interval; IV, interval variable.

Comparison of time for a task

Ten studies measured time to complete a task as the outcome.32–41 All of them expressed the measure in seconds for the task, but tasks were rather different: simple tasks like touching a target with a grasper39 or cutting a shape,31 or more complex like a part of a procedure (place a stitch with two ties).33 For this reason, the heterogeneity was high (I2=79%). Three studies reported a significant larger effect size for high level of resemblance36 40 41; two studies reported the superiority of a low level33 35; all the other studies could not show any significant difference. The pooled effect size (figure 3) was small and did not show any significant advantage of a simulation with a high over a low level of physical resemblance (−0.14, 95% CI −0.54 to 0.27).
Figure 3

Forest plot of the comparison between simulations with a high versus low level of physical resemblance for the outcome time. The pooled effect size is represented by the black diamond.

Forest plot of the comparison between simulations with a high versus low level of physical resemblance for the outcome time. The pooled effect size is represented by the black diamond.

Discussion

Our result suggests that the use of a higher or lower level of physical resemblance in simulation does not have a different effect on learning of the simple tasks that constitute the basic surgical skills. We found the same effect both on skill, measured with a score, and on the time needed to complete a task. This finding is coherent with the starting hypothesis we did, derived from learning theories. The way in which ‘the environment feature reproduced in the simulation matches the ‘‘real world’’ feature’42 is important from a cognitive point of view, because learning is enhanced by a simulation environment that recalls the real professional environment in which knowledge and skill will be used.22 Nevertheless, the skills reported in the studies we summarised in this meta-analysis were only psychomotor skills and basic technical elements, part of a more complex procedure; hence, it is not surprising that we could not find a significant difference in effectiveness between simulations based on a different level of physical resemblance. Beyond the obvious consideration on the balance between cost and result, an unnecessary level of resemblance, specially at the beginning of a technical training, can produce a cognitive overload that is detrimental for an effective learning.43 This is a possible explanation for the finding of the studies that reported a better outcome for a low level of resemblance. On the contrary, the role of a simulation based on a high level of physical resemblance is fundamental in transferring basic skills to a full procedure, as shown by many studies at the residency level.16–20 Hence, we are not arguing that the simulation of surgical procedures with advanced and immersive technology should not be used, but that the devices and the context should always be aligned with the expected learning outcomes and with the level of expertise of the trainee. The main limit of this meta-analysis is the rather small sample size of the considered studies. The pooled sample size was of 373 students; the calculations were done on a higher number of measures, and this could have introduced a bias because the same student was considered more than once. Nevertheless, most studies were coherent with the pooled effect size and with the conclusion of no difference in the comparison we did. Inconsistency among studies was from moderate to high, and this is why we adopted a random effect model that should compensate the heterogeneity. Despite these limits, we believe that this meta-analysis offers a relevant and reliable indication on the use of simulation in the training of basic surgical skill for undergraduate students.

Conclusion

Simulation with a low level of physical resemblance and the related low cost is a viable option for teaching and learning basic surgical skill at the undergraduate level. Further research should address the identification of the other elements that can contribute to a successful simulation, like peer feedback and self-regulated learning.
  42 in total

1.  Comparison of training on two laparoscopic simulators and assessment of skills transfer to surgical performance.

Authors:  Patricia L Youngblood; Sakti Srivastava; Myriam Curet; Wm Leroy Heinrichs; Parvati Dev; Sherry M Wren
Journal:  J Am Coll Surg       Date:  2005-04       Impact factor: 6.113

2.  The teaching of surgery in the undergraduate curriculum--reforms and results.

Authors:  R A Agha; A Papanikitas; M Baum; I S Benjamin
Journal:  Int J Surg       Date:  2004       Impact factor: 6.071

3.  The Tuning Project for Medicine--learning outcomes for undergraduate medical education in Europe.

Authors:  Allan Cumming; Michael Ross
Journal:  Med Teach       Date:  2007-09       Impact factor: 3.650

4.  A comparison of unguided vs guided case-based instruction on the surgery clerkship.

Authors:  William J Adamas-Rappaport; Amy L Waer; Mary K Teeple; Martin A Benjamin; Evan S Glazer; Jesse Sozanski; Diane Poskus; Evan Ong
Journal:  J Surg Educ       Date:  2012-07-26       Impact factor: 2.891

5.  Conventional Laparoscopic vs Robotic Training: Which is Better for Naive Users? A Randomized Prospective Crossover Study.

Authors:  Syed Omar Hassan; Jaimin Dudhia; Labiq H Syed; Kalpesh Patel; Maham Farshidpour; Steven C Cunningham; Gopal C Kowdley
Journal:  J Surg Educ       Date:  2015-02-14       Impact factor: 2.891

6.  Stages and transitions in medical education around the world: clarifying structures and terminology.

Authors:  Marjo Wijnen-Meijer; William Burdick; Lonneke Alofs; Chantalle Burgers; Olle ten Cate
Journal:  Med Teach       Date:  2013-01-29       Impact factor: 3.650

Review 7.  Virtual reality training in laparoscopic surgery: A systematic review & meta-analysis.

Authors:  Medhat Alaker; Greg R Wynn; Tan Arulampalam
Journal:  Int J Surg       Date:  2016-03-15       Impact factor: 6.071

Review 8.  The Effectiveness of Medical Simulation in Teaching Medical Students Critical Care Medicine: A Systematic Review and Meta-Analysis.

Authors:  Matthew David Beal; John Kinnear; Caroline Rachael Anderson; Thomas David Martin; Rachel Wamboldt; Lee Hooper
Journal:  Simul Healthc       Date:  2017-04       Impact factor: 1.929

Review 9.  Applying Educational Theory to Simulation-Based Training and Assessment in Surgery.

Authors:  Sheila W Chauvin
Journal:  Surg Clin North Am       Date:  2015-06-24       Impact factor: 2.741

Review 10.  A systematic review of low-cost laparoscopic simulators.

Authors:  Mimi M Li; Joseph George
Journal:  Surg Endosc       Date:  2016-05-18       Impact factor: 4.584

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.