Danielle Hart1, William Bond2, Jeffrey N Siegelman3, Daniel Miller4, Michael Cassara5, Lisa Barker6, Shilo Anders7, James Ahn8, Hubert Huang9, Christopher Strother10, Joshua Hui11. 1. Emergency Medicine, Hennepin County Medical Center, University of Minnesota Medical School, Minneapolis, MN. 2. Department of Emergency Medicine, Lehigh Valley Health Network, Allentown, PA. 3. Department of Emergency Medicine, Emory University, Atlanta, GA. 4. Department of Emergency Medicine, University of Iowa, Iowa City, IA. 5. Department of Emergency Medicine, Hofstra University North Shore Long Island Jewish SOM, Northwell Health Center, Lake Success, NY. 6. Department of Emergency Medicine, University of Illinois College of Medicine at Peoria, Peoria, IL. 7. Department of Anesthesiology, Vanderbilt University, Nashville, TN. 8. Department of Emergency Medicine, University of Chicago, Chicago, IL. 9. Division of Education, Lehigh Valley Health Network, Allentown, PA. 10. Emergency Medicine, Mount Sinai Hospital, New York, NY. 11. Department of Emergency Medicine, Kaiser Permanente, Los Angeles Medical Center, Los Angeles, CA.
Abstract
OBJECTIVES: All residency programs in the United States are required to report their residents' progress on the milestones to the Accreditation Council for Graduate Medical Education (ACGME) biannually. Since the development and institution of this competency-based assessment framework, residency programs have been attempting to ascertain the best ways to assess resident performance on these metrics. Simulation was recommended by the ACGME as one method of assessment for many of the milestone subcompetencies. We developed three simulation scenarios with scenario-specific milestone-based assessment tools. We aimed to gather validity evidence for this tool. METHODS: We conducted a prospective observational study to investigate the validity evidence for three mannequin-based simulation scenarios for assessing individual residents on emergency medicine (EM) milestones. The subcompetencies (i.e., patient care [PC]1, PC2, PC3) included were identified via a modified Delphi technique using a group of experienced EM simulationists. The scenario-specific checklist (CL) items were designed based on the individual milestone items within each EM subcompetency chosen for assessment and reviewed by experienced EM simulationists. Two independent live raters who were EM faculty at the respective study sites scored each scenario following brief rater training. The inter-rater reliability (IRR) of the assessment tool was determined by measuring intraclass correlation coefficient (ICC) for the sum of the CL items as well as the global rating scales (GRSs) for each scenario. Comparing GRS and CL scores between various postgraduate year (PGY) levels was performed with analysis of variance. RESULTS: Eight subcompetencies were chosen to assess with three simulation cases, using 118 subjects. Evidence of test content, internal structure, response process, and relations with other variables were found. The ICCs for the sum of the CL items and the GRSs were >0.8 for all cases, with one exception (clinical management GRS = 0.74 in sepsis case). The sum of CL items and GRSs (p < 0.05) discriminated between PGY levels on all cases. However, when the specific CL items were mapped back to milestones in various proficiency levels, the milestones in the higher proficiency levels (level 3 [L3] and 4 [L4]) did not often discriminate between various PGY levels. L3 milestone items discriminated between PGY levels on five of 12 occasions they were assessed, and L4 items discriminated only two of 12 times they were assessed. CONCLUSION: Three simulation cases with scenario-specific assessment tools allowed evaluation of EM residents on proficiency L1 to L4 within eight of the EM milestone subcompetencies. Evidence of test content, internal structure, response process, and relations with other variables were found. Good to excellent IRR and the ability to discriminate between various PGY levels was found for both the sum of CL items and the GRSs. However, there was a lack of a positive relationship between advancing PGY level and the completion of higher-level milestone items (L3 and L4).
OBJECTIVES: All residency programs in the United States are required to report their residents' progress on the milestones to the Accreditation Council for Graduate Medical Education (ACGME) biannually. Since the development and institution of this competency-based assessment framework, residency programs have been attempting to ascertain the best ways to assess resident performance on these metrics. Simulation was recommended by the ACGME as one method of assessment for many of the milestone subcompetencies. We developed three simulation scenarios with scenario-specific milestone-based assessment tools. We aimed to gather validity evidence for this tool. METHODS: We conducted a prospective observational study to investigate the validity evidence for three mannequin-based simulation scenarios for assessing individual residents on emergency medicine (EM) milestones. The subcompetencies (i.e., patient care [PC]1, PC2, PC3) included were identified via a modified Delphi technique using a group of experienced EM simulationists. The scenario-specific checklist (CL) items were designed based on the individual milestone items within each EM subcompetency chosen for assessment and reviewed by experienced EM simulationists. Two independent live raters who were EM faculty at the respective study sites scored each scenario following brief rater training. The inter-rater reliability (IRR) of the assessment tool was determined by measuring intraclass correlation coefficient (ICC) for the sum of the CL items as well as the global rating scales (GRSs) for each scenario. Comparing GRS and CL scores between various postgraduate year (PGY) levels was performed with analysis of variance. RESULTS: Eight subcompetencies were chosen to assess with three simulation cases, using 118 subjects. Evidence of test content, internal structure, response process, and relations with other variables were found. The ICCs for the sum of the CL items and the GRSs were >0.8 for all cases, with one exception (clinical management GRS = 0.74 in sepsis case). The sum of CL items and GRSs (p < 0.05) discriminated between PGY levels on all cases. However, when the specific CL items were mapped back to milestones in various proficiency levels, the milestones in the higher proficiency levels (level 3 [L3] and 4 [L4]) did not often discriminate between various PGY levels. L3 milestone items discriminated between PGY levels on five of 12 occasions they were assessed, and L4 items discriminated only two of 12 times they were assessed. CONCLUSION: Three simulation cases with scenario-specific assessment tools allowed evaluation of EM residents on proficiency L1 to L4 within eight of the EM milestone subcompetencies. Evidence of test content, internal structure, response process, and relations with other variables were found. Good to excellent IRR and the ability to discriminate between various PGY levels was found for both the sum of CL items and the GRSs. However, there was a lack of a positive relationship between advancing PGY level and the completion of higher-level milestone items (L3 and L4).
Authors: Suzanne Bentley; Stephanie N Stapleton; Phillip C Moschella; Jessica M Ray; Shana M Zucker; Jessica Hernandez; Elizabeth D Rosenman; Ambrose H Wong Journal: AEM Educ Train Date: 2019-11-27
Authors: Nur-Ain Nadir; Danielle Hart; Michael Cassara; Joan Noelker; Tiffany Moadel; Miriam Kulkarni; Christopher S Sampson; Suzanne Bentley; Neel K Naik; Jessica Hernandez; Sara M Krzyzaniak; Steven Lai; Gregory Podolej; Christopher Strother Journal: West J Emerg Med Date: 2018-11-20
Authors: Danielle Hart; Douglas Franzen; Michael Beeson; Rahul Bhat; Miriam Kulkarni; Lorraine Thibodeau; Moshe Weizberg; Susan Promes Journal: West J Emerg Med Date: 2018-11-30