Literature DB >> 33655095

The Relationship between In-Training Examinations and Simulation Assessments in an Emergency Medicine Residency Program.

Sheri L Clarke1, Ali Eydgahi2.   

Abstract

CONTEXT: Historically, the assessment of medical resident knowledge and skill has occurred through annual in-training examinations and faculty observation during real patient encounters. To improve patient care and the learning environment, medical educators have begun creating simulation experiences for medical residents to perform procedures without using real patients. However, simulation curricula and accompanying assessment techniques have not been standardized. Using a longitudinal record review, in-training examination scores were compared to annualized simulation assessment scores to see if there is any relationship between the assessment types.
METHODS: This project was a retrospective eight-year study from a single residency program. The scores were collected from 102 resident academic records from 2007 to 2015 for the annual American Board of Emergency Medicine in-training examination and the resident's annual simulation assessment. Complete data sets were analyzed to determine if a relationship exists between the assessment methods. Then the averages were compared for only the first three years for all students and for all four years for osteopathic residents as they have an additional fourth year of training.
RESULTS: This study showed a lack of relationship between the two assessment types when reviewing three years of data. When the fourth year data is considered, there was a significant relationship between the assessment types.
CONCLUSIONS: The performance scores for both types of assessment provide independent information on the resident progress in training. Therefore, they should both be reviewed and considered to appropriately measure the resident's performance. The significance of the fourth year of training for osteopathic residents requires further study.

Entities:  

Keywords:  assessment; emergency medicine; in-training examination; simulation

Year:  2016        PMID: 33655095      PMCID: PMC7746104     

Source DB:  PubMed          Journal:  Spartan Med Res J        ISSN: 2474-7629


INTRODUCTION

Assessment of medical residents’ knowledge and skill level is often done through subjective evaluations by faculty and by using objective standardized written examinations. In traditional residency training, clinical training occurs with faculty observation of real patient encounters. While this practice may be an effective educational technique for training the next generation of physicians, it puts patients at risk. To improve patient care and the learning environment, medical educators need to expand beyond traditional training and assessment methods. Simulation allows residents to practice and improve their technical skills while working on their cognitive development in a safe and non-threatening environment. Residency programs across the country are already using hybrid training models in which both simulation and traditional training are employed. Annual in-training examinations are the historical way to assess trainee medical knowledge, whereas simulation is the newest assessment technique that encompasses both medical knowledge and clinical skills. It is necessary to determine whether these assessment tools equally identify competency or provide different independent scores that should both be considered in the assessment of competency for each trainee. There is a large amount of literature that address standardized assessments in multiple medical specialties, but no literature was found that addresses simulation scores and their relationship to standardized written examinations. A summary of the most relevant literature is provided in Table 1. In Table 1, numbers represent corresponding papers listed in the references section, * denotes strong positive relationship, + shows relationship, - represents no relationship, and # indicates that poor performance on one assessment is predictive of poor performance on the other. As indicated in Table 1, the impact of the use of simulation technology for assessment has not been sufficiently explored. This study provided an analysis of the in-training examination scores and the annualized simulation scores to determine if there is any relationship between these assessment methods. The focus of the study was only on emergency medicine residents in a single program.

METHODS

This project was a retrospective study using a quantitative research method. The focus of this project was on a dually accredited residency program, which consists of osteopathic and allopathic residents. The residency program has an average of 34 residents with 10 new residents added to the program each year. This study included a sample of 102 individual residents from 2007 to 2015. The study was designed as the census of a single residency program in Lansing, Michigan to ensure that all residents in the selected sample had received the same training, used the same trainers and simulation center for all of their training, and had similar patient experiences. To improve the sample size, the study was longitudinal, with data collected for each resident in the program over an eight-year period. The institutional review board for the sponsoring hospital approved the study in affiliation with the university where this was part of a larger project for a doctoral student dissertation. Test scores were collected from the residents’ academic records for the annual American Board of Emergency Medicine (ABEM) in-training examination and the residency program’s annual simulation assessment. The score identified for each individual was the percentage of questions answered correctly with 100 being the maximum possible score. The in-training examination is valid, as it is a national standardized assessment that was designed to assess a resident’s knowledge of learning objectives set by the ABEM. The scoring for the examination is reliable, as it is a standardized examination administered by the ABEM. The simulation scores were collected using evaluation tools that were developed by the residency program faculty eight years ago. Multiple simulation scenarios and multiple assessments of similar scenarios have been combined to create an annual simulation score for each year in the program. Each of these scores were listed as a percentage of correct answers with a maximum of 100 points. The evaluation scores were assigned by one of five faculty members. The faculty has demonstrated high intra-rater and inter-rater reliability over the past eight years (unpublished data). The faculty member that moderates the simulation session scores the resident. All scores and video of the simulation are reviewed and verified by a second faculty member prior to being entered into the database. The simulation assessments are presumed to be valid because they were developed by board certified attending physicians in Emergency Medicine to assess the ABEM learning objectives. The scores are presumed to be reliable because they represent eight years of data collection, where each resident has been evaluated with the same tools through multiple observations over multiple years. In order to review and analyze the assessment methods for a relationship, the data points were separated into subsets of data. These data were identified by year of training using post graduate year (PGY). Because this was a dually accredited program, osteopathic residents who were dually enrolled in the American Osteopathic Association (AOA) and Accreditation Council for Graduate Medical Education (ACGME) had four years of data, while all other residents enrolled only in the ACGME program had three years of data. Therefore, differences in the training years were also considered by examining only the first three years of training, as well as the full data set for the osteopathic residents. Data was identified as In-Training Examination (ITE-1, ITE-2, ITE-3, ITE-4), Simulation (Sim-1, Sim-2, Sim-3, Sim-4), Average of all scores (AvgITE and AvgSim), and the first three year average of all scores (AvgITEx3 and AvgSimx3). Canonical correlation was determined to be the appropriate method for analysis as it uses correlation coefficients and weighted sums for all potential interactions to determine significance of relationships between all data subsets in a single analysis. For this type of study, using canonical correlation has several benefits over using multiple regression. It allows the researcher to review relationships with fewer calculations, but it also decreases the risk of Type I error by decreasing the number of regression equations required for analysis. Variables can be either metric or nonmetric and must have at least 10 measurements per subset in order to have an acceptable sample size. It is important to note that the correlation method does not support claims of cause and effect. It just determines whether or not the variables have a relationship. In order to infer causality, further experimental studies would need to be completed. In this study, the scores of all participants were collected to create data sets for each individual trainee. Complete data sets were analyzed in StatGraphics Software, using canonical correlation to determine if a relationship exists between the assessment methods. The averages were then compared using two sample comparisons in the statistical software.

RESULTS

The simulation scores (Sim1, Sim2, Sim3, and Sim4) and in-training examination scores (ITE1, ITE2, ITE3, and ITE4) were reviewed using canonical correlation analysis. There were 14 identified complete cases within this data. Four reviews of the variables were completed with a P-value of 0.5 and higher as shown in Table 2. This was interpreted as having no statistically significant relationship between the data sets for DO residents. A second analysis using simulation scores (Sim1, Sim2, and Sim3) and in-training examination scores (ITE1, ITE2, and ITE3) was performed to compare only the first three years of scores to make sure non-osteopathic residents are included in the analysis. There were 50 identified complete cases within this data. Three reviews of the variables revealed a P-value of 0.29 and higher as demonstrated in Table 3. This can be interpreted as having no statistically significant relationship between the data sets. The average Simulation score (AvgSim) and the average in-training examination score (AvgITE) were analyzed using two sample comparison methods to determine if they were significantly different. This is a comparison of all four years of data points. A summary of the data comparison is presented in Table 4. The report showed a Standard Skewness for AvgSim of -2.94, which indicates non-normal distribution and that comparisons based on standard deviation may not be valid. This indication of non-normal distribution led to a further analysis using Mann-Whitney U-test and Kolmogorov-Smirnov test as shown in Table 5. The Mann-Whitney U-test provides a way to compare the medians of the data sets on ordinal data. In this test a P-value of 0.02 indicated a statistically significant difference between the medians at a 95% confidence level. Then the samples were run through a Kolmogorov-Smirnov test to compare the distributions of the two samples. A P-value of 0.036 indicated a statistically significant difference between the two distributions at a 95% confidence level. These findings mean that the samples are not from similar groups, confirming that there is no relationship between the two groups, but not confirming or denying a relationship between assessment types. P-value of less than 0.05 means that data sets are significantly different at 95% confidence level. Next, the three year average Simulation score (AvgSimx3) and average in-training examination score (AvgITEx3) were reviewed to be consistent with the three year curriculum of the allopathic and international medical graduate residents. The three year data is only inclusive of complete data sets for the first three years of simulation and in-training examinations. A summary of the data comparison is shown in Table 4. The report showed a similar Standard Skewness for AvgSimx3 of 2.28, which indicates non-normal distribution and can invalidate comparisons based on standard deviation. This led to a comparison of medians using the Mann-Whitney U-test. In this test, a P-value of 0.142 indicated there is no statistically significant difference between the medians at a 95% confidence level. Then, the samples were run through a Kolmogorov-Smirnov test to compare the distributions of the two samples. A P-value of 0.09 indicated no statistically significant difference between the two distributions at a 95% confidence level. It should be noted that the three year average had a very different comparison result from the four year average as demonstrated in Table 5.

DISCUSSION

For this single residency program, the data suggest there is no relationship between the performance of residents on the ABEM in-training examination and the program’s proprietary annual simulation assessment. However, when considering the first three year average scores, the samples are not significantly different, indicating a relationship between the scores. It appears that the first three years of training have similar assessment scores and the two assessment types are similar in nature. When the fourth year of training is added for the osteopathic residents, the scores are significantly different and no longer are related. The osteopathic residents’ scores are significantly different from those of the other residents, due to an additional year of training. There has been a long discussion in the emergency medicine field as to whether the training should be three or four years. This additional year appears to make a significant difference in the relationship between the two assessment types. This could be important information for educators that are making decisions on whether emergency medicine training should be three or four years long.

CONCLUSION

This study showed a lack of relationship between the two assessment types of in-training examination and the annual simulation assessment when reviewing three years of data. The fourth year data for the osteopathic residents raised further questions. When it was considered in the analysis, it showed that there were significant differences between the osteopathic residents and all other residents. When only the first three years of training was reviewed, there appeared to be no real difference between the different medical school types. This raises a question as to why the fourth year data would make such a difference in the analysis. The authors believe that this can be explained by the fact that an additional year of data at the highest level of training is increasing the average scores for the osteopathic residents. If the resident was not providing higher scores in the fourth year, there would not be such a difference in the three and four year averages. Therefore, the performance scores for both types of assessment should be independently reviewed and considered to appropriately measure the resident’s performance. When the fourth year osteopathic data is considered, there is suddenly a significant relationship between the assessment types. This research was limited to a single residency program over eight years of data collection. It was also limited to a specific simulation process that a single program has developed and implemented. Further research opportunities would include reviewing the same data from another program or multiple programs that have either similar or different simulation assessments. In order to do further investigation on this topic, it may be beneficial to review individual resident performance in these assessments. Historically, those that do well on written examinations are thought to be more successful residents. Many times, board eligibility examination scores are used to filter applications when applying for residency. However, there are many people that have difficulty with written examinations, but excel in their field. Further research could follow individual residents to determine if those that demonstrate more knowledge on written examinations are the best at applying their knowledge in simulation. The study has provided new information on the need to consider simulation assessment as an independent metric when reviewing resident performance. Many educators assume that those that do well in medical knowledge also do well in the application of that knowledge. This study has indicated that a direct relationship between the two assessment types does not exist in emergency medicine for this single residency program’s curriculum. Another implication of this study, and any further research developed from it, is that it may assist in determining the best length of training for emergency medicine residency programs. Currently, the AOA requires a four year training program with an internship year and three years of emergency medicine training. The ACGME allows for either a three or four year length of program. With the unification of the AOA and ACGME accreditation systems, emergency medicine programs throughout the country may make significant changes in the training program length. For this particular dually accredited program, the plan is to eliminate the additional year of training for osteopathic residents as it unifies the accreditation of the programs. Does that additional year of training provide invaluable education, or are those physicians comparable to those with three years of training that spend the fourth year as an attending physician? That is the next question to answer. In order to answer that question, it would require a controlled study to analyze the simulation and in-training examination scores of the fourth year residents against the scores of the three year trained first year attending physicians. This would require a randomized sampling of residents and recent graduates across the country. It would require a significant investment in a standardized simulation assessment for all participants and a partnership with the American Board of Emergency Medicine in order to communicate with the attending physicians and collect the data necessary for comparison.

Conflict of Interest

The authors declare no conflict of interest.
Table 1. Relationships Identified in the Literature for Different Types of Assessments
COMLEX-1COMLEX-2IN-TRAININGBOARD CERTIFICATIONPERFORMANCE IN RESIDENCYBOARD REVIEW/CONFERENCESSIMULATION
USMLE-17+ 15* 16* 18+ 19+ 20+ 5- 22- 9# 11# 21#15* 8+ 10+12+ 17- 
USMLE-2 7+16* 18* 20* 5+ 19+ 22- 11# 21# 17- 
USMLE-3  16* 18+   
IN-TRAINING  16*15* 8+ 10+ 13+12+6- 14-
CLINICAL PRODUCTIVITY  13-13-  
FACULTY ASSESSMENT  4- 4+ 
INDIVIDUAL EDUC PLAN   23+  
SIMULATION     
Table 2. Canonical Correlation Analysis for Simulation and In-Training Examination Data Sets (All four years of data)
n=14CanonicalWilks
NumberEigenvalueCorrelationLambdaChi-SquareD.F.P-Value
10.5302550.7281860.24545111.9396160.7481
20.2994420.5472130.522525.5172790.7871
30.2141920.4628090.7458632.4923140.6460
40.05083270.2254610.9491670.44344710.5055
Table 3. Canonical Correlation Analysis for Simulation and In-Training Examination Data Sets (Three years only)
n=50 CanonicalWilks
NumberEigenvalueCorrelationLambdaChi-SquareD.F.P-Value
10.1593080.3991340.79023310.71290.2960
20.05850460.2418770.9399792.8163440.5890
30.001610420.040130.998390.073333210.7865
Table 4. Summary Statistics for Average Simulation and In-Training Examination Scores
All scores collectedComplete data sets only
Avg SimAvg ITEAvg Sim x3Avg ITE x3
Count (n)941025164
Average68.8871.6169.971.4
Standard deviation9.825.755.184.98
Coeff. of variation14.3 %8.0%7.40%6.97%
Minimum34.055.059.361.3
Maximum92.086.089.384.7
Range58.031.030.023.3
Stnd. skewness-2.9390.1952.2771.312
Stnd. kurtosis5.5620.1374.679-0.025
Table 5. Median and Distribution Analysis for Average Simulation and In-Training Examination Scores from Table 4
Median AnalysisDistribution Analysis
Mann-Whitney U-testKolmogorov-Smirnov test
P-valueSignificantly DifferentP-valueSignificantly Different
Avg Sim / Avg ITE0.0212178Yes0.0364329Yes
(All scores collected) (n=51)
Avg Sim x3 / Avg ITE x30.142311No0.0998673No
(Complete data sets for first three years only) (n=64)

P-value of less than 0.05 means that data sets are significantly different at 95% confidence level.

  16 in total

1.  Predictive measures of a resident's performance on written Orthopaedic Board scores.

Authors:  Bradley W Dyrstad; David Pope; Joseph C Milbrandt; Ryan T Beck; Anita L Weinhoeft; Osaretin B Idusuyi
Journal:  Iowa Orthop J       Date:  2011

2.  Predicting performance on the American Board of Surgery qualifying and certifying examinations: a multi-institutional study.

Authors:  Christian de Virgilio; Arezou Yaghoubian; Amy Kaji; J Craig Collins; Karen Deveney; Matthew Dolich; David Easter; O Joe Hines; Steven Katz; Terrence Liu; Ahmed Mahmoud; Marc L Melcher; Steven Parks; Mark Reeves; Ali Salim; Lynette Scherer; Danny Takanishi; Kenneth Waxman
Journal:  Arch Surg       Date:  2010-09

3.  Do scores of the USMLE Step 1 and OITE correlate with the ABOS Part I certifying examination?: a multicenter study.

Authors:  Paul J Dougherty; Norman Walter; Peter Schilling; Soheil Najibi; Harry Herkowitz
Journal:  Clin Orthop Relat Res       Date:  2010-03-30       Impact factor: 4.176

4.  Are United States Medical Licensing Exam Step 1 and 2 scores valid measures for postgraduate medical residency selection decisions?

Authors:  William C McGaghie; Elaine R Cohen; Diane B Wayne
Journal:  Acad Med       Date:  2011-01       Impact factor: 6.893

5.  Using United States Medical Licensing Examination(®) (USMLE) examination results to predict later in-training examination performance among general surgery residents.

Authors:  Darrell R Spurlock; Charles Holden; Thomas Hartranft
Journal:  J Surg Educ       Date:  2010-11-05       Impact factor: 2.891

6.  USMLE step 1 scores as a significant predictor of future board passage in pediatrics.

Authors:  Quimby E McCaskill; Jim J Kirk; Dawn M Barata; Peter S Wludyka; Elisa A Zenni; Thomas T Chiu
Journal:  Ambul Pediatr       Date:  2007 Mar-Apr

7.  Relationship between performance on part I of the American Board of Orthopaedic Surgery Certifying Examination and Scores on USMLE Steps 1 and 2.

Authors:  David B Swanson; Amy Sawhill; Kathleen Z Holtzman; S Deniz Bucak; Carol Morrison; Shepard Hurwitz; G Paul DeRosa
Journal:  Acad Med       Date:  2009-10       Impact factor: 6.893

Review 8.  Success in orthopaedic training: resident selection and predictors of quality performance.

Authors:  Kenneth A Egol; Jason Collins; Joseph D Zuckerman
Journal:  J Am Acad Orthop Surg       Date:  2011-02       Impact factor: 3.020

9.  Outcome measures for emergency medicine residency graduates: do measures of academic and clinical performance during residency training correlate with American Board of Emergency Medicine test performance?

Authors:  Richard C Frederick; John W Hafner; Timothy J Schaefer; Jean C Aldag
Journal:  Acad Emerg Med       Date:  2011-10       Impact factor: 3.451

10.  A comprehensive medical simulation education curriculum for emergency medicine residents.

Authors:  Emily S Binstadt; Ron M Walls; Benjamin A White; Eric S Nadel; James K Takayesu; Tobias D Barker; Stephen J Nelson; Charles N Pozner
Journal:  Ann Emerg Med       Date:  2006-12-11       Impact factor: 5.721

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.