| Literature DB >> 35238786 |
Raffaello Furlan1,2, Mauro Gatti3, Roberto Mene4, Dana Shiffer1, Chiara Marchiori5, Alessandro Giaj Levra1, Vincenzo Saturnino3, Enrico Brunetta1,2, Franca Dipaola1,2.
Abstract
BACKGROUND: Virtual patient simulators (VPSs) log all users' actions, thereby enabling the creation of a multidimensional representation of students' medical knowledge. This representation can be used to create metrics providing teachers with valuable learning information.Entities:
Keywords: clinical diagnostic reasoning; learning analytics; medical education; medical knowledge; natural language processing; virtual patient simulator
Year: 2022 PMID: 35238786 PMCID: PMC8931645 DOI: 10.2196/24372
Source DB: PubMed Journal: JMIR Med Educ ISSN: 2369-3762
Section metric description.
| Section | Sensitivity metric | Precision metric | Section metric description |
| Input scenario | Percentage of DFsa identified out of all the DFs present in the input scenario | Percentage of DFs identified in the text out of all the text selections performed by the student | Performance in identifying DFs present in the input scenario without selecting nonrelevant text |
| Anamnesis | Percentage of relevant anamnestic questions identified out of all the relevant anamnestic questions present in the simulation | Percentage of relevant anamnestic questions out of all the questions asked by the student | Performance in asking all the relevant questions without asking superfluous questions |
| Physical examination | Percentage of relevant physical examinations performed out of all the relevant physical examinations present in the simulation | Percentage of relevant physical examinations performed out of all the physical examinations performed by the student | Performance in carrying out all the relevant physical examinations without carrying out superfluous physical examinations |
| Medical test | Percentage of relevant medical tests requested out of all the relevant medical tests present in the simulation | Percentage of relevant medical tests requested out of all the medical tests requested by the student | Performance in requesting all the relevant medical tests without asking for superfluous medical tests |
| DHb | Percentage of reasonable DHs identified out of all the reasonable DHs present in the simulation | Percentage of reasonable DHs identified out of all the DHs formulated by the student | Performance in identifying all the reasonable DHs without formulating inappropriate DHs |
| BAc | Percentage of BA mappings correctly executed on the first attempt out of the total number of BA mappings present in the simulation | Percentage of BA mappings correctly executed on the first attempt out of the total number of BA mappings executed by the student | Performance in identifying the correct DF–DH relationships (increase, neutral, and decrease) on the first attempt |
| Final diagnosis | Percentage of correct diagnoses identified by the student out of the total number of correct diagnoses present in the simulation | Percentage of correct diagnoses identified by the student out of the total number of diagnoses (correct and incorrect) formulated by the student | Performance in identifying the correct final diagnoses |
aDF: diagnostic factor.
bDH: diagnostic hypothesis.
cBA: binary analysis.
Figure 1Class overall performance scores during tests 1 and 2 as shown by histogram bar distribution. During test 2, the presence of bars on the left side points to the existence of students characterized by a weaker overall performance compared with the rest of the class. The range of each bar is 0.05.
Figure 2Relationship between collection and analytical scores during test 1 (April 12) and test 2 (May 21). Each dot represents the performance of a single student. The ideal (maximal) performance score corresponds to 1.0. The dashed line indicates the median of the overall scores of the class. Note that students 202025 and 202041 (arrows) reached a similar overall score (0.63) in different ways. Student 202041 performed worse in the data collection exercise (collection rank 12 and analysis rank 5), whereas student 202025 performed poorly in the data analysis exercise (collection rank 3 and analysis rank 11) compared with the class results.
Figure 3Radar graphs of the top- and bottom-performing students and average class results in each exercise section during test 1. Graphs enabled the comparison between the scores of the different exercise sections of the simulation as obtained by the top (continuous line) and bottom (long dashed line and grey area) performers and by the class (short dashed line). Note that the top-performing student scored consistently better than the average of the class on all tasks except the history-taking exercise. In contrast, the bottom performer scored less in every exercise except the anamnesis. The 2 students could be given individualized advice by teachers to overcome each specific weakness. The results refer to test 2. AN: anamnesis; BA: binary analysis; HY: hypothesis generation; MT: medical tests; PE: physical examination; RS: results; SC: scenario.
Figure 4Relationship between individual overall scores and corresponding methodological scores obtained during test 2. The arrow indicates the students who scored weakly as far as the clinical methodology is concerned, although the overall score was acceptable. Therefore, this student is specifically lacking in their way of addressing that diagnosis and needs ad hoc teacher’s advice.
Figure 5Critical diagnostic acts and expected execution path during test 2. The most likely diagnosis in that simulation was cholecystitis, and the key actions the user was expected to perform from the start (S) were previously set to be (1) palpation of the abdomen (right upper quadrant; P), (2) check for the Murphy sign (M), (3) request for abdomen ultrasonography (U), and (4) final diagnosis (D), corresponding to the PMUD pathway (thin yellow arrow). Each arrow represents a different execution flow. The width of the arrow is proportional to the number of students who followed that flow. Note that, of the 36 students, only 3 (8%) executed all 3 crucial diagnostic acts in the expected order, whereas 16 (44%) reached the correct final diagnosis without performing a physical exam, and 8 (22%) gave priority to abdomen sonography.
Results of the Spearman rank correlation test between the hematology examination score and the 4 main Hepius metrics.
| Metric | Correlation index | |
| Overall score | 0.2867 | .22 |
| Collection score | 0.2786 | .23 |
| Analytical score | −0.0404 | .87 |
| Methodological score | 0.0836 | .73 |