Literature DB >> 26594293

Emergency Medicine Residents Consistently Rate Themselves Higher than Attending Assessments on ACGME Milestones.

Katja Goldflam1, Jessica Bod1, David Della-Giustina1, Alina Tsyrulnik1.   

Abstract

INTRODUCTION: In 2012 the Accreditation Council for Graduate Medical Education (ACGME) introduced the Next Accreditation System (NAS), which implemented milestones to assess the competency of residents and fellows. While attending evaluation and feedback is crucial for resident development, perhaps equally important is a resident's self-assessment. If a resident does not accurately self-assess, clinical and professional progress may be compromised. The objective of our study was to compare emergency medicine (EM) resident milestone evaluation by EM faculty with the same resident's self-assessment.
METHODS: This is an observational, cross-sectional study that was performed at an academic, four-year EM residency program. Twenty-five randomly chosen residents completed milestone self-assessment using eight ACGME sub-competencies deemed by residency leadership as representative of core EM principles. These residents were also evaluated by 20 faculty members. The milestone levels were evaluated on a nine-point scale. We calculated the average difference between resident self-ratings and faculty ratings, and used sample t-tests to determine statistical significance of the difference in scores.
RESULTS: Eighteen residents evaluated themselves. Each resident was assessed by an average of 16 attendings (min=10, max=20). Residents gave themselves statistically significant higher milestone ratings than attendings did for each sub-competency examined (p<0.0001).
CONCLUSION: Residents over-estimated their abilities in every sub-competency assessed. This underscores the importance of feedback and assessment transparency. More attention needs to be paid to methods by which residency leadership can make residents' self-perception of their clinical ability more congruent with that of their teachers and evaluators. The major limitation of our study is small sample size of both residents and attendings.

Entities:  

Mesh:

Year:  2015        PMID: 26594293      PMCID: PMC4651597          DOI: 10.5811/westjem.2015.8.27247

Source DB:  PubMed          Journal:  West J Emerg Med        ISSN: 1936-900X


INTRODUCTION

In 2012 the Accreditation Council for Graduate Medical Education (ACGME) introduced the Next Accreditation System (NAS), which implemented milestones to assess the abilities and progress of residents. Each milestone is a significant, progressive, competency-based point in the development of a resident. These milestones evaluate accomplishments that identify specialty-specific knowledge, skills, attitudes and behaviors that can be used as outcome measures within the general competencies.1,2 Emergency medicine (EM) has developed 23 sub-competencies, with five milestone levels within each. Residents are expected to progress through levels of proficiency as they complete their training.2,3 Attending evaluation and feedback is crucial for resident development. However, at least as equally important is a resident’s self-assessment. This is because feedback from others is often interpreted and integrated through the framework of a learner’s self-assessment.4,5 Learners use an amalgam of self-assessment and feedback to generate actionable goals for improvement.6 If a resident does not accurately self-assess, clinical and professional progress may be compromised. A resident who is unable to accurately judge his or her own abilities may fail to achieve the necessary skills to be a safe and effective physician. In other words, failure to acknowledge deficiencies may lead to a failure to correct them. Attending physicians working at academic centers are used as the benchmark in assessing a resident’s abilities as a physician. To date, no study has compared resident self-assessment to attending assessment using the standardized framework of the ACGME milestones.

Goals of this Investigation

Our study used the framework of the ACGME milestones to compare EM resident evaluation by EM faculty with the same residents’ self-assessments.

METHODS

This study is an observational, cross-sectional study performed at an academic EM residency. A human investigation committee (HIC) exemption was granted by the institutional IRB. All residents from EM post graduate year (PGY) 1 through 4 level were included in the study, with the exception of the single resident who helped to conduct the study. Twenty-five residents were chosen using a random number generator to participate in the study. The remaining residents were omitted due to time limitations on attendings filling out the forms and concerns that too large a number of evaluations would be prohibitive to attending willingness to participate in the study. These residents completed self-assessments of milestone levels using eight ACGME sub-competencies that were chosen as representative of core EM principles by residency leadership consensus. Moreover, residency leadership agreed that a large group of attending evaluators would likely be able to comment on these, more familiar, sub-competencies for the majority of residents. The residency leadership consensus consisted of the residency program director and associate program directors. These included Emergency Stabilization (PC1), History and Physical (PC2), Diagnostic Studies (PC3), Diagnosis (PC4), Disposition (PC7), Communication (ICS1), Multi-Tasking (PC8), and Team Management (ICS2). These same residents were also evaluated by 20 faculty members using identical milestones. Faculty members have contact with residents in various settings, which include clinical shifts, simulation laboratory, and in small-group teaching sessions. Faculty members were able to opt out of assessing any resident whom they felt they could not evaluate due to limited interaction. The sub-competencies were evaluated on a nine-point scale, which reflects the rubric published by the ACGME (Figure). No advanced training or instruction was provided regarding the utilization of the ACGME milestones. No other evaluation tools were provided to faculty when they were asked to assign a score.
Figure

Sample Accreditation Council for Graduate Medical Education milestone used to assess competency of emergency medicine residents and fellows.

We calculated the average difference between resident self-ratings and faculty ratings. Sample t-tests were used to determine the statistical significance of the difference in scores. We carried out mixed models analyses to determine if there were any significant interactions between the rater type (self vs. attending) and program year. For each program year, we calculated and compared the difference in the least square means between residents and their attending raters to the overall difference in least square means for each sub-competency.

RESULTS

Eighteen of the 25 residents surveyed completed the evaluation. Each resident was assessed by an average of 16 attendings (min=10, max=20). Residents gave themselves higher milestone ratings than attendings did for each of the eight sub-competencies evaluated (Table 1). The mean difference in score for each sub-competency was close to one point, with the exception of “Team Management,” which was 0.5 points. For seven out of eight sub-competencies, the difference in resident milestone self-assessment score and attending milestone assessment score was statistically significant (p<0.05). The one sub-competency where statistical significance was not reached was “Team Management” (p=0.09).
Table 1

Comparison of all residents’ post graduate years 1–4 self-rating to attending rating.

Evaluation constructSelfRater(s)Estimated difference ± standard error95% CLp
Communication6.68 ± 0.335.54 ± 0.101.14 ± 0.32(0.49, 1.78)0.0006
Diagnosis6.77 ± 0.305.60 ± 0.561.16 ± 0.31(0.08, 2.34)0.0002
Diagnostic studies6.89 ± 0.295.62 ± 0.081.26 ± 0.29(0.68, 1.84)<0.0001
Disposition6.54 ± 0.315.52 ± 0.091.01 ± 0.31(0.39, 1.63)0.0015
Emergency stabilization6.22 ± 0.305.51 ± 0.080.70 ± 0.30(0.10, 1.30)0.0212
History and physical6.95 ± 0.335.72 ± 0.081.23 ± 0.33(0.57, 1.89)0.0003
Multi-tasking6.80 ± 0.335.48 ± 0.081.31 ± 0.33(0.65, 1.97)0.0001
Team management5.99 ± 0.315.47 ± 0.100.52 ± 0.30(−0.08, 1.13)0.0902
Mixed model analysis showed statistically significant differences between self-ratings and attending ratings in most sub-competencies for the PGY 1 and 3 cohorts (Table 2 and Table 3). The PGY 2 cohort had fewer differences across sub-competencies, with statistically significant differences in only three sub-competencies (Table 4). For PGY 4, self and attending ratings did not significantly differ in any sub-competency (Table 5).
Table 2

Comparison of post graduate year 1 self-rating to attending rating.

Evaluation constructSelfRater(s)Estimated difference ± standard error95% CLp
Communication6.16 ± 0.604.14 ± 0.212.01 ± 0.60(0.81, 3.21)0.001
Diagnosis5.17 ± 0.563.95 ± 0.161.21 ± 0.57(0.09, 2.34)0.0348
Diagnostic studies6.56 ± 0.544.03 ± 0.172.52 ± 0.55(1.45, 3.60)<0.0001
Disposition5.35 ± 0.583.92 ± 0.191.42 ± 0.58(0.27, 2.58)0.0152
Emergency stabilization4.95 ± 0.563.69 ± 0.181.25 ± 0.56(0.14, 2.36)0.0265
History & physical6.56 ± 0.614.12 ± 0.182.44 ± 0.62(1.21, 3.66)0.0001
Multi-tasking5.17 ± 0.613.77 ± 0.171.39 ± 0.62(0.17, 2.62)0.0254
Team management4.92 ± 0.574.07 ± 0.210.84 ± 0.57(−0.27, 1.97)0.1392
Table 3

Comparison of post graduate year 3 self-rating to attending rating.

Evaluation constructSelfRater(s)Estimated difference ± standard error95% CLp
Communication6.84 ± 0.555.79 ± 0.181.05 ± 0.55(−0.03, 2.13)0.0573
Diagnosis7.50 ± 0.515.91 ± 0.141.59 ± 0.52(0.56, 2.61)0.0024
Diagnostic studies7.17 ± 0.495.98 ± 0.151.19 ± 0.49(0.21, 2.16)0.0165
Disposition6.51 ± 0.535.87 ± 0.160.63 ± 0.53(−0.40, 1.68)0.2291
Emergency stabilization6.99 ± 0.515.89 ± 0.161.10 ± 0.51(0.10, 2.11)0.0311
History & physical7.66 ± 0.556.08 ± 0.151.58 ± 0.56(0.47, 2.68)0.0053
Multi-tasking7.49 ± 0.555.72 ± 0.151.76 ± 0.56(0.65, 2.87)0.0019
Team management6.31 ± 0.525.68 ± 0.180.62 ± 0.51(−0.39, 1.64)0.2272
Table 4

Comparison of post graduate year 2 self-rating to attending rating.

Evaluation constructSelfRater(s)Estimated difference ± standard error95% CLP
Communication6.14 ± 0.675.07 ± 0.201.07 ± 0.67(−0.24, 2.39)0.1113
Diagnosis6.70 ± 0.635.19 ± 0.151.50 ± 0.63(0.25, 2.75)0.0181
Diagnostic studies6.43 ± 0.605.15 ± 0.161.28 ± 0.60(0.09, 2.46)0.0346
Disposition6.91 ± 0.655.07 ± 0.181.83 ± 0.64(0.56, 3.11)0.0048
Emergency stabilization5.87 ± 0.625.07 ± 0.170.79 ± 0.62(−0.43, 2.02)0.2043
History & physical5.92 ± 0.685.20 ± 0.170.72 ± 0.68(−0.62, 2.07)0.2934
Multi-tasking6.19 ± 0.685.10 ± 0.161.08 ± 0.68(−0.26, 2.44)0.1153
Team management6.12 ± 0.645.03 ± 0.201.09 ± 0.63(−0.15, 2.33)0.0863
Table 5

Comparison of post graduate year 4 self-rating to attending rating.

Evaluation constructSelfRater(s)Estimated difference ± standard error95% CLP
Communication7.58 ± 0.787.15 ± 0.190.43 ± 0.77(−1.09, 1.95)0.58
Diagnosis7.71 ± 0.727.35 ± 0.150.35 ± 0.73(−1.08, 1.79)0.626
Diagnostic studies7.39 ± 0.697.34 ± 0.160.04 ± 0.69(−1.31, 1.41)0.9434
Disposition7.38 ± 0.747.23 ± 0.180.15 ± 0.74(−1.31, 1.62)0.8399
Emergency stabilization7.07 ± 0.727.40 ± 0.17−0.33 ± 0.71(−1.74, 1.08)0.643
History & physical7.67 ± 0.787.47 ± 0.160.19 ± 0.79(−1.36, 1.75)0.8054
Multi-tasking8.33 ± 0.787.32 ± 0.161.01 ± 0.79(−0.54, 2.57)0.203
Team management6.59 ± 0.737.06 ± 0.20−0.46 ± 0.73(−1.90, 0.97)0.5271

DISCUSSION

Our study found that residents (combined PGY1 through PGY4) consistently rated themselves as more proficient for each sub-competency than did their attending evaluators. This is consistent with prior data showing that physician self-assessment typically does not correlate with external measures of performance.7 Although self-assessment may be inaccurate, it is important for evaluators to consider learner self-image when giving feedback. This feedback will undoubtedly be interpreted by the learner through a filter of his/her own perception.8 For example, feedback from an attending that is lower than a learner feels he/she attained, may be rejected by that learner who believes he or she has reached a higher level of proficiency. This could negatively impact the development and growth of that learner. Our study illustrates that milestone-based assessment remains subject to these considerations. This suggests that educators must be cognizant of residents’ self-assessments when formulating and delivering feedback. Our subgroup analysis included small sample sizes; more work with larger sample sizes is necessary to determine if program year does indeed have an effect on agreement between resident and attending assessment. Within this context, our data showed that differences between self-assessment and attending assessment may be affected by program year. Unlike the results for PGY 1 through 3, self and attending ratings for PGY 4 did not differ significantly on any sub-competency. These results suggest that in PGY 4, self and attending ratings converge and are quite similar. It is important to note that these p values were not adjusted for multiple comparisons and should therefore be interpreted as only part of further exploratory analyses. Taking the results of our study into consideration, the finding that residents perceive themselves as more capable than they are rated by attendings would be relevant to discussions in Clinical Competency Committee (CCC) meetings. Residents’ perception of their skills would be important in grading them on sub-competencies that deal with “practice-based learning and improvement.” Although difficult to put into practice, perhaps resident self-evaluations should be included in their “residency portfolios” and compared to the CCC rating of that resident to ensure that as the resident moves through the program self-perception is not significantly different from that of his/her evaluators.

LIMITATIONS

The major limitation of this study is the small sample size of attendings and residents evaluated. Our self-assessment response rate was 70%. Self-assessment was not compulsory, as participation in research was voluntary per our HIC. It is not known if those who did not respond were different demographically or in PGY year, as the study personnel was blinded to the identities of the residents assessed. Another limitation is the varying levels of familiarity with the milestones among the residents and attendings surveyed. In addition, due to the nature of EM shift work, attendings have different frequencies of interactions with residents, which may introduce bias into their assessments. As the study was done at an academic institution, some faculty members do have less clinical time in the department than others. Although faculty members were permitted to opt out of assessing a resident with whom they had limited experience, we acknowledge that the frequency and types of faculty-resident interactions assessed may vary widely. In addition, the residents studied were at different levels of their training; this may have influenced their self-ratings. For example, some of the residents surveyed were close to graduation, a circumstance that may inflate their self-assessments. Although our data suggest that PGY 4 residents’ and attending evaluations may be similar, our interpretation is limited by the small number of representatives in each class; thus, more investigation is required to determine if there is a difference between classes in their ability to self-assess accurately. A larger sample of residents assessed may allow for more detailed sub-group analysis by PGY year. In addition, a larger sample size would also allow for more detailed analysis of high and low performers and their ability to self-assess, as had been demonstrated in the past. This study relies on the assumption that attending ratings are more accurate than resident self-rating, the validity of which may need further investigation.4,7 Perhaps most importantly, milestones are a relatively new assessment tool with very few studies evaluating their validity.1,3,9

CONCLUSION

Residents over-estimated their abilities in each of eight sub-competencies assessed. This underscores the importance of feedback and assessment transparency. More attention needs to be paid to methods by which residency leadership can make residents’ clinical ability self-perception more congruent with that of their teachers and evaluators.
  9 in total

1.  The next GME accreditation system--rationale and benefits.

Authors:  Thomas J Nasca; Ingrid Philibert; Timothy Brigham; Timothy C Flynn
Journal:  N Engl J Med       Date:  2012-02-22       Impact factor: 91.245

2.  The processes and dimensions of informed self-assessment: a conceptual model.

Authors:  Joan Sargeant; Heather Armson; Ben Chesluk; Timothy Dornan; Kevin Eva; Eric Holmboe; Jocelyn Lockyer; Elaine Loney; Karen Mann; Cees van der Vleuten
Journal:  Acad Med       Date:  2010-07       Impact factor: 6.893

3.  The emergency medicine milestones: a validation study.

Authors:  Robert C Korte; Michael S Beeson; Chad M Russ; Wallace A Carter; Earl J Reisdorff
Journal:  Acad Emerg Med       Date:  2013-07       Impact factor: 3.451

Review 4.  The development of the emergency medicine milestones.

Authors:  Michael S Beeson; Wallace A Carter; Theodore A Christopher; Jonathan W Heidt; James H Jones; Lynne E Meyer; Susan B Promes; Kevin G Rodgers; Philip H Shayne; Susan R Swing; Mary Jo Wagner
Journal:  Acad Emerg Med       Date:  2013-06-19       Impact factor: 3.451

5.  Emergency medicine residents' self-assessments play a critical role when receiving feedback.

Authors:  Richard Bounds; Colleen Bush; Amish Aghera; Nestor Rodriguez; R Brent Stansfield; Sally A Santen
Journal:  Acad Emerg Med       Date:  2013-10       Impact factor: 3.451

6.  Practicing emergency physicians report performing well on most emergency medicine milestones.

Authors:  Timothy C Peck; Nicole Dubosh; Carlo Rosen; Carrie Tibbles; Jennifer Pope; Jonathan Fisher
Journal:  J Emerg Med       Date:  2014-07-08       Impact factor: 1.484

Review 7.  Accuracy of physician self-assessment compared with observed measures of competence: a systematic review.

Authors:  David A Davis; Paul E Mazmanian; Michael Fordis; R Van Harrison; Kevin E Thorpe; Laure Perrier
Journal:  JAMA       Date:  2006-09-06       Impact factor: 56.272

Review 8.  Systematic review of the literature on assessment, feedback and physicians' clinical performance: BEME Guide No. 7.

Authors:  Jon Veloski; James R Boex; Margaret J Grasberger; Adam Evans; Daniel B Wolfson
Journal:  Med Teach       Date:  2006-03       Impact factor: 3.650

9.  Factors influencing responsiveness to feedback: on the interplay between fear, confidence, and reasoning processes.

Authors:  Kevin W Eva; Heather Armson; Eric Holmboe; Jocelyn Lockyer; Elaine Loney; Karen Mann; Joan Sargeant
Journal:  Adv Health Sci Educ Theory Pract       Date:  2011-04-06       Impact factor: 3.853

  9 in total
  10 in total

1.  Implementation of a pilot novel objective peer comparison evaluation system in an emergency medicine residency program.

Authors:  Kraftin E Schreyer; Megan E Healy; Zachary Repanshek; Wayne A Satz; Jacob W Ufberg
Journal:  AEM Educ Train       Date:  2022-03-31

2.  Integrating self-assessment into feedback for emergency medicine residents.

Authors:  Jenna Thomas; Benjamin Sandefur; James Colletti; Aidan Mullan; James Homme
Journal:  AEM Educ Train       Date:  2022-02-01

3.  Progressive Surgical Autonomy in a Plastic Surgery Resident Clinic.

Authors:  Kristopher M Day; Jillian K Scott; Lani Gao; Tara M Lee; Jimmy L Waldrop; Larry A Sargent; J Woody Kennedy; Jason P Rehm; Mark A Brzezienski
Journal:  Plast Reconstr Surg Glob Open       Date:  2017-05-04

4.  Are All Competencies Equal in the Eyes of Residents? A Multicenter Study of Emergency Medicine Residents' Interest in Feedback.

Authors:  Suzanne Bentley; Kevin Hu; Anne Messman; Tiffany Moadel; Sorabh Khandelwal; Heather Streich; Joan Noelker
Journal:  West J Emerg Med       Date:  2016-12-15

5.  Use of Emergency Department Pharmacists in Emergency Medicine Resident Milestone Assessment.

Authors:  Starr-Mar'ee C Bedy; Kara B Goddard; Julie A W Stilley; Christopher S Sampson
Journal:  West J Emerg Med       Date:  2018-12-05

6.  Trials and Tribulations in Implementation of the Emergency Medicine Milestones from the Frontlines.

Authors:  Alexander Y Sheng
Journal:  West J Emerg Med       Date:  2019-06-03

7.  Resident Self-Assessment and the Deficiency of Individualized Learning Plans in Our Residencies.

Authors:  David Della-Giustina; Ali Kamran; D Brian Wood; Katja Goldflam
Journal:  West J Emerg Med       Date:  2020-12-11

8.  Critical Appraisal of Emergency Medicine Educational Research: The Best Publications of 2015.

Authors:  Corey R Heitz; Wendy Coates; Susan E Farrell; Jonathan Fisher; Amy Miller Juve; Lalena M Yarris
Journal:  AEM Educ Train       Date:  2017-10-17

9.  Designing and Implementing Emergency Department Pain Management Curriculum: A Delphi Approach.

Authors:  Karolina Paziana; Edmund Timpano; Andrew Stolbach
Journal:  AEM Educ Train       Date:  2018-04-02

10.  Resident Perceptions of a Publicly Disclosed Daily Productivity Dashboard.

Authors:  Katja Goldflam; Alina Tsyrulnik; Colin Flood; Jessica Bod; Ryan F Coughlin; David Della-Giustina
Journal:  West J Emerg Med       Date:  2022-01-03
  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.