Literature DB >> 32736583

Online patient simulation training to improve clinical reasoning: a feasibility randomised controlled trial.

Ruth Plackett1, Angelos P Kassianos2, Maria Kambouri3, Natasha Kay2, Sophie Mylan4, Jenny Hopwood5, Patricia Schartau6, Shani Gray2, Jessica Timmis2, Sarah Bennett5, Chris Valerio5, Veena Rodrigues7, Emily Player7, Willie Hamilton8, Rosalind Raine2, Stephen Duffy9, Jessica Sheringham2.   

Abstract

BACKGROUND: Online patient simulations (OPS) are a novel method for teaching clinical reasoning skills to students and could contribute to reducing diagnostic errors. However, little is known about how best to implement and evaluate OPS in medical curricula. The aim of this study was to assess the feasibility, acceptability and potential effects of eCREST - the electronic Clinical Reasoning Educational Simulation Tool.
METHODS: A feasibility randomised controlled trial was conducted with final year undergraduate students from three UK medical schools in academic year 2016/2017 (cohort one) and 2017/2018 (cohort two). Student volunteers were recruited in cohort one via email and on teaching days, and in cohort two eCREST was also integrated into a relevant module in the curriculum. The intervention group received three patient cases and the control group received teaching as usual; allocation ratio was 1:1. Researchers were blind to allocation. Clinical reasoning skills were measured using a survey after 1 week and a patient case after 1 month.
RESULTS: Across schools, 264 students participated (18.2% of all eligible). Cohort two had greater uptake (183/833, 22%) than cohort one (81/621, 13%). After 1 week, 99/137 (72%) of the intervention and 86/127 (68%) of the control group remained in the study. eCREST improved students' ability to gather essential information from patients over controls (OR = 1.4; 95% CI 1.1-1.7, n = 148). Of the intervention group, most (80/98, 82%) agreed eCREST helped them to learn clinical reasoning skills.
CONCLUSIONS: eCREST was highly acceptable and improved data gathering skills that could reduce diagnostic errors. Uptake was low but improved when integrated into course delivery. A summative trial is needed to estimate effectiveness.

Entities:  

Keywords:  Clinical reasoning; Feasibility study; Medical education; Medical students; Online learning; Simulation; Virtual patient

Mesh:

Year:  2020        PMID: 32736583      PMCID: PMC7395338          DOI: 10.1186/s12909-020-02168-4

Source DB:  PubMed          Journal:  BMC Med Educ        ISSN: 1472-6920            Impact factor:   2.463


Background

Clinical reasoning — the thought processes used by clinicians during consultations to formulate appropriate questions — is essential for timely diagnosis of disease [1-4]. Providing training in clinical reasoning as early as possible in medical education could improve reasoning skills in future doctors, as it provides a scaffold for future learning, and retraining reasoning can be challenging [5, 6]. However, in undergraduate medical education, there is a lack of explicit teaching on clinical reasoning and the development and delivery of additional high quality and consistent clinical reasoning teaching potentially increases burden on faculty’s already stretched time and resources [2, 7, 8]. Online patient simulations (OPS) are a specific type of computer-based program that simulates real-life clinical scenarios and could support teaching reasoning skills [7, 9]. Theories of cognition suggest that exposure to a large number of different clinical cases via simulations could improve reasoning by restructuring and building more complex mental representations [10, 11]. Learning by experience also facilitates reflection, which helps students to retain skills [12]. OPS can be blended with traditional teaching and offers the opportunity for students to practise data gathering and make diagnoses without burdening patients [9, 13]. OPS also have pragmatic benefits; once developed they are lower in cost to deliver, can be distributed widely, completed remotely, tailored to the learner and frequently updated [9, 14]. Nevertheless, using technology-enhanced learning (TEL) may have its own limitations, such as the potential for lack of engagement from users and faculty, lack of fidelity with real patient consultations and limited TEL skills of faculty [15, 16]. There is currently little empirical evidence to support the use of OPS for assisting clinical reasoning skills teaching. The few studies conducted were not methodologically robust and difficult to interpret due to the poor validity of clinical reasoning outcome measures [17-20]. Furthermore, most previous studies had limited information on the feasibility of introducing a novel tool into a curriculum and evaluating them using a robust research method, such as a randomised controlled trial (RCT) [17, 18]. Understanding the feasibility of testing OPS using an RCT study design is necessary before a summative RCT can estimate effectiveness [21, 22]. This research aimed to inform the design of a summative evaluation of an OPS to support teaching of reasoning skills in medical schools. The development of this OPS, the Electronic Clinical Reasoning Educational Simulation Tool (eCREST), is reported elsewhere [23]. Briefly, eCREST shows three videos of patients (played by actors) presenting to their primary care physician (PCP) with respiratory problems that could be indicative of serious conditions like lung cancer. The student gathers information from the patient, while continually being prompted to review their differential diagnosis. After each case they are asked to make a final differential diagnosis and receive feedback. Patient cases were developed with a small group of real patients who co-wrote the scripts of the vignettes and helped to identify pertinent clinical and behavioural characteristics for the simulated cases [24]. This study sought to obtain evidence as to the feasibility of a trial through: identifying optimal recruitment strategies, measured by student uptake; testing the acceptability to students via student retention and feedback; testing the validity and measuring the possible effect sizes of two clinical reasoning outcome measures.

Methods

Study design and participants

A multicentre parallel feasibility RCT was conducted across three UK medical schools: A, B and C. We followed the CONSORT statement for reporting pilot or feasibility trials [25]. Eligible participants were final year undergraduate medical students. The curricula of the medical schools varied. Schools A and B implemented a traditional integrated/systems-based curriculum. School C followed a problem-based learning (PBL) curriculum. Ethical approval was gained from participating medical schools. Participants were recruited from March 2017–February 2018 in two cohorts. Cohort one was recruited after final examinations in April–July 2017, through advertisements in faculty newsletters and lecture ‘shout outs’. Cohort two was recruited prior to final examinations in October 2017–February 2018. School C students were only recruited in cohort two. Cohort two were invited to participate through the faculty online learning management platforms (e.g. Moodle), advertisements on social media, faculty newsletters, and lecture ‘shout outs’. As this was a feasibility trial, a sample size calculation was not required.

Outcomes

Feasibility and acceptability

Feasibility was measured by assessing student uptake by school and cohort. Acceptability was measured by retention rates and a survey adapted from previous studies, consisting of six statements on the perceptions of eCREST [26, 27].

Clinical reasoning outcome measures

Clinical reasoning was measured using the Flexibility in Thinking (FIT) scale of the Diagnostic Thinking Inventory (DTI), which is a self-reported measure [28]. The FIT (21 items) measures thought processes used in the diagnostic process, including the ability to generate new ideas, understand alternative outcomes and self-reflect. Higher scores on the FIT sub-scale are indicative of better clinical reasoning skills. The sub-scale has demonstrated validity to detect differences between student and professional reasoning. The internal consistency and test re-test reliability were acceptable [28, 29]. Clinical reasoning was also measured using an observed measure of clinical reasoning by using data from an additional eCREST patient case that students received 1 month after baseline. This measure comprised indicators of three cognitive biases that eCREST sought to influence: the unpacking principle, confirmation bias and anchoring. These were identified by previous clinical reasoning research [24, 30, 31]. The unpacking principle refers to the tendency to not elicit the necessary information to make an informed judgement. Confirmation bias is when a clinician only seeks information to confirm their hypothesis. Anchoring occurs when clinicians stick to an initial hypothesis despite contradictory information [32]. eCREST prompts students to reflect throughout a consultation and provides feedback that enables them to reflect on their performance afterwards [33]. By reflecting, students would be more likely to attend to evidence inconsistent with their hypotheses and consider alternatives, thereby reducing the chance of confirmation bias and anchoring. Reflection also encourages students to explore their hypotheses thoroughly, ensuring that they elicit relevant information from patients, reducing the effect of the unpacking principle [33, 34]. The observed measure assessed ‘essential information identified’ by measuring the proportion of essential questions and examinations asked, out of all possible essential examinations and questions identified by experts. This aimed to detect the influence of the unpacking principle on reasoning, as it captured whether the students elicited enough essential information to make an appropriate decision. The ‘relevance of history taking’ was measured by assessing the proportion of all relevant questions and examinations asked, out of the total questions and examinations asked by the student. This aimed to detect susceptibility to confirmation bias by capturing whether they sought relevant information. Finally, it measured ‘flexibility in diagnoses’ by counting the number of times students changed their diagnosis. This reflected how susceptible students were to anchoring, by measuring their willingness to change their initial differential diagnosis. All measures were developed by RP and three clinicians (PS, SG & JT). The content validity of the observed measure of clinical reasoning was tested with two clinicians (SM, JH).

Diagnostic choice

Diagnostic choice was captured in the additional patient case. Selection of the most important diagnosis that the student should not have missed was used to assess how well the observed measure of reasoning predicted diagnostic choice.

Knowledge

Relevant medical knowledge was measured by 12 single best answer multiple choice questions (MCQs). We hypothesised that greater knowledge is associated with better clinical reasoning skills, consistent with the literature [4, 35]. The MCQs were developed by clinicians (NK, SM, JH & PS) in consultation with other clinicians.

Procedure

The trial procedure is outlined in Fig. 1, which shows how and when data from participants were collected. To address ethical concerns the information sheet made it clear to students that: participation in the trial was voluntary, they could withdraw at any stage, participation would not impact upon their summative assessments and only anonymised aggregate data would be shared. Students who provided written consent online were allocated to intervention or control groups using simple randomisation. Researchers were blind to allocation, completed by a computer algorithm. Randomisation was not precisely 1:1, as five students were mistakenly automatically allocated to the intervention group. The intervention group received three video patient cases in eCREST, all presenting with respiratory or related symptoms to their primary care physician [23]. The control group received no additional intervention and received teaching as usual. To address concerns that students in the control group may be disadvantaged by not having access to eCREST, we ensured that the control group had access to eCREST at the end of the trial.
Fig. 1

Flow diagram illustrating study procedure

Flow diagram illustrating study procedure

Data analysis

Uptake was calculated as the percentage of students who registered out of the total number of eligible students. Retention was calculated as the percentage of students who completed T1 and T2 follow-up assessments out of all registered. Acceptability was measured by calculating the percentage of students who agreed with each statement on the acceptability questionnaire. Uptake, retention and acceptability were compared between schools and cohorts using chi-squared tests.

Clinical reasoning outcomes

Validity and reliability

Internal consistency of the self-reported clinical reasoning measure was assessed using Cronbach’s alpha. Construct validity of the self-reported and observed clinical reasoning measures was assessed by correlating the reasoning and knowledge outcomes, using Spearman’s rank correlation coefficient. To estimate the predictive validity of the clinical reasoning measures, the self-reported measure and observed measure of clinical reasoning were correlated with diagnostic choice. The analyses were undertaken for the aggregated dataset then separately for the intervention and control groups.

Effect sizes

Independent t-tests were used to compare mean self-reported clinical reasoning scores between intervention and control groups at T1 and T2. A mixed factorial ANOVA was used to assess change in self-reported clinical reasoning over time, between groups and interaction effects. Logistic regression analyses were conducted to assess the ‘essential information identified’ and the ‘relevance of history taking’. These outcomes were proportional data, so were transformed by calculating the log odds of the outcomes [36, 37]. Group allocation was the only predictor variable in each model, as knowledge did not significantly differ between the groups at baseline. A multinomial logistic analysis was carried out to assess ‘Flexibility in diagnoses’. A complete case analysis was undertaken, such that those students who had missing data were excluded from analysis. Analyses were conducted using Stata Version 15, with p ≤ 0.05 considered statistically significant [38].

Results

Feasibility and acceptability

Across the three UK medical schools, 264 students participated (18.2% of all eligible, Fig. 2). Recruitment was greater for students in cohort two (n = 183/833, 22%) than those in cohort one (n = 81/621, 13%). Uptake was slightly greater at school B (n = 136/610, 22%), followed by A (n = 112/696, 16%) and C (n = 16/148, 11%). Uptake was similar at schools A (n = 44//336, 13%) and B (n = 37/285, 13%) in cohort one. However, uptake was greater at school B (n = 99/325, 31%) than A (n = 68/360, 19%) and C (n = 16/148, 11%) in cohort two. Participant characteristics are shown in Table 1 and no significant differences between the intervention and control group were observed.
Fig. 2

Flow diagram showing how participants progress through the feasibility RCT

Table 1

Participant characteristics at baseline

Intervention groupControl groupP value
Age (Years): n (%)
 20–224 (2.9)1 (0.8)
 23–2473 (53.3)79 (62.2)
 25–2639 (28.5)29 (22.8)
 27–2811 (8.0)10 (7.9)
  > 2910 (7.3)8 (6.3)0.49
 Total n137127
Gender: n (%)
 Female64 (46.7)58 (45.7)
 Male73 (53.3)69 (54.3)0.87
 Total n137127
 Self-reported clinical reasoning skills at baseline: mean (SD)83.1 (9.6)83.5 (8.8)0.75
 Total n122118
 Knowledge at baseline: mean (SD)9.2 (1.8)9.3 (1.6)0.22
 Total n125126

Notes: Baseline n varied for demographics, self-reported clinical reasoning skills and knowledge as some students did not complete all information. Results are presented as number and percentage of students or mean and standard deviation for each group. Comparisons were made using t-tests for means and Chi-squared tests for percentages. P Value less than 0.05 was considered significant

Flow diagram showing how participants progress through the feasibility RCT Participant characteristics at baseline Notes: Baseline n varied for demographics, self-reported clinical reasoning skills and knowledge as some students did not complete all information. Results are presented as number and percentage of students or mean and standard deviation for each group. Comparisons were made using t-tests for means and Chi-squared tests for percentages. P Value less than 0.05 was considered significant There was no significant difference detected in retention between the intervention and control groups 1 week after baseline, 72 and 68% respectively (χ2 (1) = 0.65, p = 0.42), or after 1 month, 57 and 55% respectively (χ2 (1) = 0.34, p = 0.56, Fig. 2). There was no significant difference found in the proportion of students at each school who stayed in the study 1 week after baseline. However, there was significantly poorer retention at school A after 1 month (n = 47/112, 42%) than at school B (n = 83/136, 61%) and C (n = 10/16, 63%), χ2 (2) = 9.58, p = 0.008. Those in cohort one were significantly less likely to stay in the study one-week post baseline (n = 45/81, 56%) than those in cohort two (n = 140/183, 77%), χ2 (1) = 11.75, p = 0.001. This was also observed one-month post baseline (n = 29/81, 36% and n = 111/183, 61% respectively), χ2 (1) = 13.92, p = 0.000. Most students (> 80%) agreed that eCREST helped them learn clinical reasoning skills and that they would use it again without incentives (Table 2). There were no significant differences detected between the schools. However, those in cohort two were significantly more likely than cohort one to agree that: eCREST helped to improve their clinical reasoning skills (87.7% vs 64.0%), χ2 (2) = 7.5, n = 98, p = .024); eCREST enhanced their overall learning (93.2% vs 64.0%), χ2 (2) = 13.7, n = 98, p = .001) and that they would use eCREST again without an incentive (97.3% vs 52.0%), χ2 (2) = 31.8, n = 98, p = .000).
Table 2

Intervention group medical student responses to the acceptability survey

Strongly agree/ AgreeNeither agree or disagreeStrongly disagree/ Disagree
Statementn(%)n(%)n(%)
It was easy to navigate through eCREST96/98 (98)1/98 (1)1/98 (1)
The level of difficulty of the material was appropriate95/98 (97)3/98 (3)0/98 (0)
eCREST should be used to supplement traditional teaching88/98 (90)9/98 (9)1/98 (1)
eCREST helped me to learn clinical reasoning skills to apply to clinical work80/98 (82)15/98 (15)3/98 (3)
Overall, using eCREST enhanced my learning84/98 (86)13/98 (13)1/98 (1)
I would use eCREST in the future without an incentive84/98 (86)10/98 (10)4/98 (4)

Notes: results are taken from across all 3 schools. 98 students in the intervention group completed the acceptability survey at Time 1

Intervention group medical student responses to the acceptability survey Notes: results are taken from across all 3 schools. 98 students in the intervention group completed the acceptability survey at Time 1

Clinical reasoning outcomes

Validity

The internal consistency of the self-reported clinical reasoning measure was adequate (Cronbach’s α = 0.66). Correlations between self-reported and observed clinical reasoning outcome measures, and knowledge and clinical outcomes are shown in Table 3. There was a mostly positive but non-significant correlation between the self-reported clinical reasoning measure and the observed clinical reasoning measure. The self-reported clinical reasoning measure had a weak but significant positive correlation with knowledge for aggregated data (r = 0.13, p = 0.037, n = 240). The observed clinical reasoning measure was positively but not significantly correlated with knowledge. The self-reported clinical reasoning measure at baseline and the observed clinical reasoning measure were positively but not significantly correlated with identification of the most serious diagnosis.
Table 3

Mean scores and correlations between measures of clinical reasoning, knowledge and diagnostic choice

Reasoning measureGroupaNMean Scores (SD)Correlation coefficients
Self-reported clinical reasoning at baselinebKnowledge at baselinecDiagnostic choiced
Self-reported clinical reasoning skillsrseprsprsp
 BaselineI12283.1 (9.6)0.170.057−0.010.927
C11883.5 (8.8)0.100.2700.090.463
All24083.3 (9.2)0.140.037*0.040.663
 Time 1fI9984.1 (10.3)
C8682.4 (9.0)
All18583.3 (9.8)
 Time 2gI7584.4 (9.8)
C6582.0 (9.4)
All14083.3 (9.7)
Observed clinical reasoning skillsh
 Essential information identifiediI7861.6% (17.6)−0.010.9570.090.4230.090.435
C7053.3% (15.8)0.030.7980.220.0740.130.303
All14857.7% (17.2)0.010.8650.150.780.120.144
 Relevance of history takingjI7881.4% (10.5)0.060.5780.060.6230.120.304
C7084.6% (10.6)0.020.8930.110.3560.080.518
All14882.9% (10.6)0.030.7040.070.4150.100.251
 Flexibility in diagnoseskI783.2 (1.0)0.010.9190.170.1370.010.933
C703.0 (1.0)0.090.4390.040.7720.230.054
All1483.1 (1.0)0.050.5420.110.1860.120.137

aIntervention (I) and Control group (C)

bMeasured using the Flexibility in Thinking scale (Bordage et al. 1990)

cMeasured by 12 multiple choice respiratory medicine questions

dMeasured by whether the most important diagnosis was selected for a patient case

ers denotes Spearman’s rank correlation coefficient

fTime 1 = one week after registration

gTime 2 = one month registration

hMeasured by performance on a patient case delivered by eCREST to all students

iPercentage of essential information from gathered from patient case out of possible essential information available

jPercentage of relevant information from gathered from patient case out of all information student gathered

kNumber of times changed diagnosis

* indicates p ≤ 0.05

Mean scores and correlations between measures of clinical reasoning, knowledge and diagnostic choice aIntervention (I) and Control group (C) bMeasured using the Flexibility in Thinking scale (Bordage et al. 1990) cMeasured by 12 multiple choice respiratory medicine questions dMeasured by whether the most important diagnosis was selected for a patient case ers denotes Spearman’s rank correlation coefficient fTime 1 = one week after registration gTime 2 = one month registration hMeasured by performance on a patient case delivered by eCREST to all students iPercentage of essential information from gathered from patient case out of possible essential information available jPercentage of relevant information from gathered from patient case out of all information student gathered kNumber of times changed diagnosis * indicates p ≤ 0.05

Effect sizes

The intervention group had non-significantly higher self-reported clinical reasoning skills than the control group at Time 1 (84.1 vs 82.4, p = 0.26) and Time 2 (84.4 vs 82.0, p = 0.15). There was no significant effect of group allocation, (F (1)=0.00, p = 0.97, n = 136) time, (F (2)=0.01, p = 0.99, n = 136) or interaction between group allocation and time, F (2)=0.48, p = 0.62, n = 136. Table 4 shows logistic regression analyses comparing observed clinical reasoning skills between the intervention and control groups. The intervention group identified significantly more essential information than the control group (62% vs 53%). The control group sought more relevant information than the intervention group (85% vs 81%) but this difference was not significant. Students in both groups changed their diagnoses at least twice. The intervention group changed their diagnoses more often than controls, but the difference was not statistically significant.
Table 4

Logistic regression analyses comparing clinical reasoning skills between intervention and control groups

Observed clinical reasoning skillsaTrial groupnMean (SD)Odds Ratio (95% CI)P value
Essential information identifiedbIntervention7861.6% (17.6)
Control7053.3% (15.8)
1.40 (1.12, 1.75)0.003*
Relevance of history takingcIntervention7881.4% (10.5)
Control7084.6% (10.6)
0.79 (0.62, 1.01)0.064d
Flexibility in diagnoseseIntervention783.2 (1.0)
Control703.0 (1.0)
2 (base)
31.48 (0.68, 3.24)0.323
41.63 (0.68, 3.92)0.270
52.46 (0.55, 11.00)0.239
61.77 (0.07, 20.76)0.887f

aMeasured by students’ performance on a patient case delivered by eCREST

bPercentage of essential information from gathered from patient case out of possible essential information available

cPercentage of relevant information from gathered from patient case out of all information student gathered

dχ2 (1) =3.44

eNumber of times changed diagnosis

fχ2 (4) =2.24, p = 0.692

*indicates p ≤ 0.05

Logistic regression analyses comparing clinical reasoning skills between intervention and control groups aMeasured by students’ performance on a patient case delivered by eCREST bPercentage of essential information from gathered from patient case out of possible essential information available cPercentage of relevant information from gathered from patient case out of all information student gathered dχ2 (1) =3.44 eNumber of times changed diagnosis fχ2 (4) =2.24, p = 0.692 *indicates p ≤ 0.05

Discussion

This feasibility trial of eCREST demonstrated that optimal recruitment and retention was achieved when the tool was integrated into curricula, as seen in the greater uptake for cohort two than cohort one. eCREST was also highly acceptable to students, suggesting it would be feasible to conduct a summative trial to estimate effectiveness of OPS in medical schools, if there was course integration. Uptake, retention and acceptability were higher amongst students in cohort two than cohort one. Providing eCREST to students before exams, and advertising and integrating with students’ online learning management platforms, may have made eCREST more accessible and useful for revision. The low uptake at school C compared to the other sites was possibly due to their different curriculum design. Students at school C may have had more exposure to patient cases than A and B, as part of their PBL designed curriculum, reducing the need for simulated cases [39]. In a summative trial, recruitment efforts made for cohort two would likely yield greater uptake but uptake may vary across schools with different curricular approaches. Given the effect sizes observed in this study, we estimate that a sample size of 256 would be sufficient to detect a significant increase in the proportion of essential information identified. However, schools considering implementing OPS should be mindful that the acceptability of OPS to students could be affected by barriers to adoption at faculty level, such as insufficient technological capabilities to adapt and manage OPS and a lack of alignment of OPS content with educators’ needs [40]. To explore factors that might affect uptake we are undertaking qualitative research to understand how students’ reason and interact with eCREST and to understand from a faculty perspective how novel tools like eCREST can be implemented into curricula. A lack of validated measures to assess clinical reasoning skills has been reported in medical education literature [10, 18]. This study assessed the suitability of two potential measures. The FIT self-reported measure of clinical reasoning had some construct validity but poor predictive validity. It also may not capture actual reasoning as it is a self-reported measure. The observed measure of clinical reasoning developed for this study measured real-time thought processes involved in making decisions but had poor construct and predictive validity. Difficulties in establishing the validity of any clinical reasoning measure arise because of the subjective nature and context-dependency of clinical reasoning [10]. In future, predictive validity of observed measures may be better established by applying rubrics to several patient cases and correlating with measures of summative performance that require strong clinical reasoning, such as objective structured clinical examination (OSCE) performance on related cases. The observed clinical reasoning outcomes suggest that eCREST could reduce the effects of the unpacking principle, and confirmation and anchoring biases. eCREST helped students to elicit more information from patients on symptoms indicative of serious diseases and encouraged students to challenge their original hypotheses. Nevertheless, it is also possible that eCREST encouraged students to be less efficient when gathering information, by increasing the number of questions they asked. This may be an unfeasible approach in clinical practice given the significant time pressures clinicians face. However, given medical students’ limited level of experience, and exposure to patients and knowledge, this strategy may be appropriate when managing patients with non-specific symptoms in primary care [13].

Limitations

There was relatively low uptake in the study (18%) but the extensive demands of medical curricula often result in low uptake of additional resources. Uptake was higher in this study than some previous online learning studies that relied on medical student volunteers and the sample size was ample for the purposes of the feasibility RCT [41, 42]. This study demonstrated that some integration of eCREST into the curricula in cohort two was possible and led to greater uptake, acceptability and retention. However, as this study relied on volunteers there was a risk of selection bias. Students who took part might have been different than those who did not. A further limitation of this study is that it was a complete case analysis, which assumes data were missing at random and those who dropped out were similar to those who remained. It was not possible to follow up those who dropped out to determine whether they had different views of eCREST or different patterns of reasoning. Furthermore, the observed measure of clinical reasoning was only collected at T2 and not at baseline. It is possible that the two groups differed on this measure at baseline. However, no differences in self-reported clinical reasoning were detected between groups at baseline.

Conclusions

This feasibility RCT has illustrated the importance of integration into the course when evaluating OPS in medical education. It would be feasible to conduct a summative trial to assess the effectiveness of eCREST on medical students’ clinical reasoning skills in multiple medical schools, if it were appropriately positioned in a curriculum to benefit student learning. Further testing of the validity of using OPS as an outcome measure is needed. Nevertheless, this study provides evidence that OPS can be used to support face-to-face teaching to reduce cognitive biases, which may help future doctors in achieving timely diagnoses in primary care.
  32 in total

1.  Computer assisted learning in undergraduate medical education.

Authors:  T Greenhalgh
Journal:  BMJ       Date:  2001-01-06

2.  Framework for design and evaluation of complex interventions to improve health.

Authors:  M Campbell; R Fitzpatrick; A Haines; A L Kinmonth; P Sandercock; D Spiegelhalter; P Tyrer
Journal:  BMJ       Date:  2000-09-16

3.  Increasing medical students' engagement in public health: case studies illustrating the potential role of online learning.

Authors:  J Sheringham; A Lyon; A Jones; J Strobl; H Barratt
Journal:  J Public Health (Oxf)       Date:  2015-10-20       Impact factor: 2.341

4.  Diagnostic errors and reflective practice in medicine.

Authors:  Sílvia Mamede; Henk G Schmidt; Remy Rikers
Journal:  J Eval Clin Pract       Date:  2007-02       Impact factor: 2.431

5.  The Causes of Errors in Clinical Reasoning: Cognitive Biases, Knowledge Deficits, and Dual Process Thinking.

Authors:  Geoffrey R Norman; Sandra D Monteiro; Jonathan Sherbino; Jonathan S Ilgen; Henk G Schmidt; Silvia Mamede
Journal:  Acad Med       Date:  2017-01       Impact factor: 6.893

6.  The Role for Virtual Patients in the Future of Medical Education.

Authors:  Norman B Berman; Steven J Durning; Martin R Fischer; Soren Huwendiek; Marc M Triola
Journal:  Acad Med       Date:  2016-09       Impact factor: 6.893

7.  Variations in GPs' decisions to investigate suspected lung cancer: a factorial experiment using multimedia vignettes.

Authors:  Jessica Sheringham; Rachel Sequeira; Jonathan Myles; William Hamilton; Joe McDonnell; Judith Offman; Stephen Duffy; Rosalind Raine
Journal:  BMJ Qual Saf       Date:  2016-09-20       Impact factor: 7.035

8.  Developing and evaluating complex interventions: the new Medical Research Council guidance.

Authors:  Peter Craig; Paul Dieppe; Sally Macintyre; Susan Michie; Irwin Nazareth; Mark Petticrew
Journal:  BMJ       Date:  2008-09-29

9.  The incidence of diagnostic error in medicine.

Authors:  Mark L Graber
Journal:  BMJ Qual Saf       Date:  2013-06-15       Impact factor: 7.035

10.  Training of clinical reasoning with a Serious Game versus small-group problem-based learning: A prospective study.

Authors:  Angélina Middeke; Sven Anders; Madita Schuelper; Tobias Raupach; Nikolai Schuelper
Journal:  PLoS One       Date:  2018-09-11       Impact factor: 3.240

View more
  8 in total

1.  Planning a Collection of Virtual Patients to Train Clinical Reasoning: A Blueprint Representative of the European Population.

Authors:  Anja Mayer; Vital Da Silva Domingues; Inga Hege; Andrzej A Kononowicz; Marcos Larrosa; Begoña Martínez-Jarreta; Daloha Rodriguez-Molina; Bernardo Sousa-Pinto; Małgorzata Sudacka; Luc Morin
Journal:  Int J Environ Res Public Health       Date:  2022-05-19       Impact factor: 4.614

2.  Virtual, Augmented, and Alternate Reality in Medical Education: Socially Distanced but Fully Immersed.

Authors:  Stacey M Kassutto; Cameron Baston; Caitlin Clancy
Journal:  ATS Sch       Date:  2021-10-18

3.  Medical students' perspective on online teaching during pandemic: Experience from a Government Medical College in Uttarakhand, India.

Authors:  Pooja Sharma Kala; Naveen Thapliyal; Hari Shankar Pandey; A R Piyush; Sonam Maheshwari; Vikram Singh Chaudhary
Journal:  J Educ Health Promot       Date:  2021-12-31

4.  The effectiveness of using virtual patient educational tools to improve medical students' clinical reasoning skills: a systematic review.

Authors:  Ruth Plackett; Angelos P Kassianos; Sophie Mylan; Maria Kambouri; Rosalind Raine; Jessica Sheringham
Journal:  BMC Med Educ       Date:  2022-05-13       Impact factor: 3.263

5.  Implementing Remote Collaboration in a Virtual Patient Platform: Usability Study.

Authors:  Jan Kiesewetter; Inga Hege; Michael Sailer; Elisabeth Bauer; Claudia Schulz; Manfred Platz; Martin Adler
Journal:  JMIR Med Educ       Date:  2022-07-28

6.  The effects of neonatal resuscitation gamification program using immersive virtual reality: A quasi-experimental study.

Authors:  Sun-Yi Yang; Yun-Hee Oh
Journal:  Nurse Educ Today       Date:  2022-07-06       Impact factor: 3.906

7.  Advances in e-learning in undergraduate clinical medicine: a systematic review.

Authors:  T Delungahawatta; S S Dunne; S Hyde; L Halpenny; D McGrath; A O'Regan; C P Dunne
Journal:  BMC Med Educ       Date:  2022-10-07       Impact factor: 3.263

8.  Using Virtual Patients to Explore the Clinical Reasoning Skills of Medical Students: Mixed Methods Study.

Authors:  Ruth Plackett; Angelos P Kassianos; Jessica Timmis; Jessica Sheringham; Patricia Schartau; Maria Kambouri
Journal:  J Med Internet Res       Date:  2021-06-04       Impact factor: 5.428

  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.