Literature DB >> 35583658

Development of a novel behaviourally anchored instrument for the assessment of surgical trainees.

Tzong-Yang Pan1, Frank Piscioneri1, Cathy Owen2.   

Abstract

BACKGROUND: The Royal Australasian College of Surgeons (RACS) created its competency framework in 2003 which initially consisted of nine competencies each regarded as equally important for a practising surgeon. The JDocs Framework is aligned to these competencies and provides guidance for junior doctors working towards the Surgical Education and Training program.
METHODS: A novel assessment instrument was designed around the JDocs framework using 48 behaviourally anchored questions. The study was completed in 2020 across five public hospitals in the ACT and NSW. Participants were invited to complete the self-assessment form online.
RESULTS: Thirty-six of 59 (61%) trainees participated in the study, with 67 of 68 (98.5%) supervisors having completed the assessment form. Trainee self-rating scores were lower than that of supervisor ratings across all competencies except communication. The self-rating scores were negatively correlated with the seniority of a trainee's level in all nine competencies. The years of post-graduate experience was positively correlated with seven of the nine competencies. For gender and International Medical Graduate status, correlation was only identified for health advocacy and medical expertise. There was no correlation identified with a trainee's age.
CONCLUSION: This pilot study has provided an opportunity to explore a new assessment instrument for surgical trainees that is aligned to the RACS competency framework using behaviourally anchored questions. Looking ahead, a better understanding of this instrument will potentially be helpful in early identification of underperforming trainees in order to facilitate early intervention, or its use as a selection tool for formal training programs.
© 2022 The Authors. ANZ Journal of Surgery published by John Wiley & Sons Australia, Ltd on behalf of Royal Australasian College of Surgeons.

Entities:  

Keywords:  JDocs framework; RACS competencies; surgical trainee assessment; workplace based assessment

Mesh:

Year:  2022        PMID: 35583658      PMCID: PMC9544592          DOI: 10.1111/ans.17767

Source DB:  PubMed          Journal:  ANZ J Surg        ISSN: 1445-1433            Impact factor:   2.025


Introduction

The Royal Australasian College of Surgeons (RACS) is the leading advocate for surgical standards in Australia and New Zealand. Ten core competencies required by a surgeon have been described by the college since 2003, encompassing the domains of Medical Expertise, Judgement – Clinical Decision‐Making, Technical Expertise, Professionalism and Ethics, Health Advocacy, Communication, Collaboration and Teamwork, Management and Leadership, and Scholarship and Teaching. Each competency is deemed equally important and are assessed throughout surgical education and training by supervisors and examination boards. The JDocs Framework is aligned to these competencies and provides guidance for junior doctors working towards the Surgical Education and Training program. Trainees who are awarded Fellowship of the Royal Australasian College of Surgeons (FRACS) are recognized as competent in all of these 10 domains and considered to be qualified to practice independently as a surgeon. Workplace based assessment is increasingly being utilized to assess surgical trainees regularly through their training. These assessments not only provide an opportunity for feedback and reflection, but can also identify struggling trainees so support can be provided early. These assessments can be extended beyond supervisors and include various colleagues or even patients in the form of a 360° assessment, which has been well validated in many specialities of medicine and surgery. , , Assessment instruments can be further improved by using behaviourally anchored rating scales as they lead to more consistent ratings due to their more clinically relevant descriptors, rather than descriptors of values and principles. , The increasing acceptance of workplace‐based assessments has naturally led to the emergence of new assessment instruments. However, many of these instruments are focused on a specific skill or domain. For example, there are workplace‐based assessments which focus on ‘communication in the operating theatre’ or ‘professional behaviour’. There remains a need for a comprehensive assessment instrument that encompasses all the skills required of a surgeon and is aligned to the RACS framework. The aim of this study was to develop a novel instrument for assessment of technical and non‐technical skills for surgical trainees that is based off the nine RACS competencies utilizing the JDocs framework. The additional 10th RACS competency on Cultural competence and cultural safety, was implemented after this study was already undertaken so was not included in the design of the assessment tool (Data S1). This pilot study had three primary questions: Does this novel instrument have adequate internal reliability to be used in more comprehensive assessments? Do supervisors have a tendency to provide ratings that are higher, or lower, than trainee self‐ratings using this instrument and method of delivery? What variables of the trainee can be used to estimate their self‐ratings for each competency?

Methods

A novel instrument was designed based off the RACS nine competencies utilizing the JDocs framework. Components of the instrument were reviewed by three independent senior surgical and medical educators and refined to produce a 48‐item assessment tool covering all nine competencies. Invitation to participate in the study was sent in April 2019 to all 59 trainees working in the ACT and South‐East NSW health network, including prevocational surgical registrars, SET trainees and fellows. Background demographic data was collected which included age, gender, post‐graduate year, level of training, and whether they were an International Medical Graduate (IMG). Participants then rated themselves in all 48 items on a Likert scale. Participants were asked to nominate two surgical supervisors to assess them using an analogous instrument modified for supervisors. All collected data was de‐identified and the instrument was accessible for 12 weeks on SurveyMonkey for participants to complete. Statistical analysis of the data was completed using SPSS 25. The reliability of the data for both trainees and supervisors were calculated using the Cronbach's Alpha test. The student's t‐test (independent two‐sample) was used to determine if there was a significant difference between the trainees' self‐rating, to that of the averaged supervisor ratings. A multiple regression model using age, gender, post‐graduate year, IMG status, and level of training as the variables was performed with backwards elimination, and pairwise comparisons made to identify the degree and direction of influence each variable contributed to trainee self‐ratings.

Results

During the survey period, 36 of 59 (61%) trainees completed the assessment. Two responses were grossly incomplete and were excluded from further analysis. From 25 nominated supervisors, 67 of 68 (98.5%) assessments were fully completed, 1 was partially complete. Five of these supervisors had completed the Foundation Skills for Surgical Educators course and all had completed the Operating With Respect course. The demographic data of the trainees is presented in Table 1. The Cronbach's Alpha reliability for each competency is presented in Table 2 and shows good consistency across all nine competency ratings for both trainee and supervisor responses with the exception of health advocacy for trainees. The student's t‐test presented in Table 3 shows that for all competencies except communication, there is a statistically significant difference in self‐rating scores compared with those of supervisors.
Table 1

Demographic of trainees

Age25–2911 (32.4%)
30–3412 (35.3%)
35+11 (32.4%)
GenderFemale12 (35.3%)
Male22 (64.7%)
PGY3–411 (32.4%)
5–610 (29.4%)
7+13 (38.2%)
Level of trainingPrevocational22 (64.7%)
SET9 (26.5%)
Fellow3 (8.8%)
IMGNo29 (85.3%)
Yes5 (14.7%)
Table 2

Internal consistency

CompetencyNumberTrainees (α)NumberSupervisors (α)
Communication340.918680.827
Collaboration and teamwork340.806680.702
Management and leadership340.889680.755
Professionalism and ethics340.856660.789
Health advocacy340.447680.713
Scholarship and teaching340.904670.819
Medical expertise340.881670.867
Judgement – clinical decision making340.784670.823
Technical expertise340.832670.845
Table 3

Student's t‐test between trainee and supervisor ratings

CompetencyTrainee self‐ratingAveraged supervisor rating P‐value
Communication4.2184.3870.097
Collaboration and teamwork4.1624.3460.033
Management and leadership3.8914.2390.000
Professionalism and ethics3.8614.3080.000
Health advocacy3.8824.3460.000
Scholarship and teaching3.8124.2240.000
Medical expertise3.9604.3340.000
Judgement – clinical decision making4.1404.3400.029
Technical expertise3.9264.3510.000
Demographic of trainees Internal consistency Student's t‐test between trainee and supervisor ratings The multiple regression analysis undertaken looking at trainee self‐rating scores is shown in Table 4. Trainees with more years of post‐graduate experience were estimated to provide self‐ratings that were higher across seven of the nine competencies compared to those with less years of experience. The post‐graduate year of the trainee was identified as statistically significant in its correlation in the domains of communication, management and leadership, professionalism and ethics, scholarship and teaching, medical expertise, judgement and technical expertise. The level of training level was identified as statistically significant in its correlation with the competencies of communication, collaboration and teamwork, management and leadership, professionalism and ethics, health advocacy, scholarship and teaching, medical expertise, judgement and technical expertise. While for both gender and IMG status, correlation was only identified for health advocacy and medical expertise. There was no correlation identified for age.
Table 4

Multiple regression model for each of the nine RACS competencies

CompetencyTermsEstimate of parameter (se)χ2‐statistics df P‐value
CommunicationBaseline value4.039 (0.1066)
AcceptedPostgraduate yearPGY 3–40 (baseline)

17.55

2

<0.001

PGY 5–6+0.530 (0.1177)
PGY 7++0.779 (0.1705)
Training LevelPrevocational0 (baseline)29.392<0.001
SET−0.556 (0.1839)
Fellow−1.437 (0.2660)
RejectedGender0.05410.816
IMG0.06710.796
Age0.75520.755
Collaboration and teamworkBaseline value4.216 (0.0887)
AcceptedTraining levelPrevocational0 (baseline)

3409.1

3

<0.001
SET+0.006 (0.1387)
Fellow−0.633 (0.2402)
RejectedPostgraduate year3.04020.219
Gender0.00710.933
IMG1.16610.280
Age0.26120.878
Management and leadershipBaseline value3.740 (0.1423)
Accepted

Postgraduate year

PGY 3–40 (baseline)7.87320.020
PGY 5–6+0.424 (0.1571)
PGY 7++0.727 (0.2276)

Training Level

Prevocational0 (baseline)10.0102

0.007

SET−0.606 (0.2454)
Fellow−1.039 (0.3551)
RejectedGender2.55410.110
IMG2.63810.104
Age1.88520.390
Professionalism and ethicsBaseline value3.766 (0.1329)
Accepted

Postgraduate year

PGY 3–40 (baseline)6.7472

0.034

PGY 5–6+0.188 (0.1467)
PGY 7++0.643 (0.2124)

Training Level

Prevocational

SET

Fellow

0 (baseline)

−0.483 (0.2291)

−0.885 (0.3314)

8.00320.018
RejectedGender3.58510.058
IMG3.49110.062
Age0.2782

0.870

Health advocacyBaseline value3.814 (0.1042)
AcceptedTraining levelPrevocational0 (baseline)12.17820.002
SET−0.122 (0.1500)
Fellow−1.272 (0.3932)

Gender

Male0 (baseline)

4.575

10.032
Female+0.317 (0.1482)

IMG

No0 (baseline)5.08310.024
Yes+0.686 (0.3043)
RejectedAge0.07520.963
Postgraduate year2.32920.374
Scholarship and teachingBaseline value3.691 (0.1606)
Accepted

Postgraduate year

PGY 3–40 (baseline)6.60020.037
PGY 5–6+0.195 (0.1773)
PGY 7++0.753 (0.2568)

Training level

Prevocational0 (baseline)6.41920.040
SET−0.492 (0.2769)
Fellow−0.978 (0.4006)
RejectedGender2.17910.140
IMG1.64910.199
Age0.3032

0.860

Medical expertiseBaseline value3.588 (0.1033)
Accepted

Postgraduate year

PGY 3–40 (baseline)25.55920.000
PGY 5–6+0.551 (0.1085)
PGY 7++0.814 (0.1567)

Training level

Prevocational0 (baseline)29.59420.000
SET−0.458 (0.1657)
Fellow−1.524 (0.2942)
GenderMale0 (baseline)4.04510.044
Female+0.226 (0.1123)

IMG

No0 (baseline)4.58810.032
Yes+0.505 (0.2357)
RejectedAge0.59920.741
Judgement – clinical decision makingBaseline value3.955 (0.1099)
Accepted

Postgraduate year

PGY 3–40 (baseline)15.92020.000
PGY 5–6+0.378 (0.1214)
PGY 7++0.823 (0.1758)

Training Level

Prevocational0 (baseline)17.47120.000
SET−0.541 (0.1895)
Fellow−1.112 (0.2742)
RejectedGender0.20510.651
IMG0.19710.657
Age3.74220.154
Technical expertiseBaseline value3.455 (0.1400)
AcceptedPostgraduate YearPGY 3–40 (baseline)23.29620.000
PGY 5–6+0.768 (0.1545)
PGY 7++1.2 (0.2238)

Training level

Prevocational0 (baseline)14.47120.001
SET−0.365 (0.2413)
Fellow−1.322 (0.3492)
RejectedGender2.78710.095
IMG2.34410.126
Age1.49120.474
Multiple regression model for each of the nine RACS competencies 17.55 2 <0.001 3409.1 3 Postgraduate year Training Level 0.007 Postgraduate year 0.034 Training Level Prevocational SET Fellow 0 (baseline) −0.483 (0.2291) −0.885 (0.3314) 0.870 Gender 4.575 IMG Postgraduate year Training level 0.860 Postgraduate year Training level IMG Postgraduate year Training Level Training level The multiple regression model utilizes backward elimination. The baseline values for the estimate of parameter of each competency are presented with the estimated change in value for each subgroup.

Discussion

As boards of surgical training programs and surgical institutions continue to explore the use of work‐based assessments, it is becoming more important to understand the relationships and correlations between the instrument and trainees. This pilot of our instrument to trainees and supervisors was simple in its delivery. The response rates for both trainees (61%) and supervisors (98.5%) represent excellent participation in the survey which support the acceptability of the delivery method. The design we utilized can be easily expanded to include more supervisors, or ratings from colleagues in other work relationships such as nurses. The flexibility of this instrument means that it can be adapted to be used as a comprehensive 360° assessment, or a simple 1‐on‐1 supervisor feedback template. In this pilot study, we have demonstrated excellent internal consistency of each competency in the instrument for both trainee and supervisor ratings, the exception being poor reliability of trainee self‐ratings in health advocacy. Interestingly, supervisor ratings for this competency demonstrates good reliability. Whether this finding suggests that supervisors are truly able to provide more consistent evaluations of this competency than trainees, or are reliably providing the same rating to both questions as they have no idea how to answer, is a point of interest. Future iterations of this instrument can clarify this and improve the consistency in this competency by expanding on the current two questions to at least four–six questions to make it similar to the other competencies. In the comparison of trainee self‐ratings with averaged supervisor ratings, supervisors gave higher ratings across eight of the nine competencies. This finding is consistent with other instruments in the literature demonstrating that supervisor ratings are higher than self‐ratings. Trainees are thought to be more modest in self‐ratings, while supervisors may feel that providing low scores may result in a plethora of further inquiries about the trainee's performance and the additional work to be a deterrent. This tendency to give higher ratings is a problem as it may not accurately identify trainees who may be ‘slightly underperforming’. However, trainees who are ‘severely underperforming’ may still be given ratings far enough below the average to be identified as a struggling trainee. If a greater number of supervisors are used in this instrument, then it will improve the likelihood of identifying struggling trainees and provide the opportunity for early academic and personal support. The data shows that trainees of a higher post‐graduate year experience are estimated to rate themselves higher across seven of the nine competencies. The exceptions to this being ‘collaboration and teamwork’ and ‘health advocacy’ (Table 4). This implies that simply having more years of clinical experience does not directly result in trainees feeling more competent in these two areas. Targeted training in these two areas may be most valuable for trainees and surgical training programs can consider placing more time and emphasis on them in the future. The model reveals an interesting finding about the level of training of trainees with their self‐ratings in all nine of the competencies. Prevocational trainees are expected to rate themselves the highest, but as they progressed along their career into a SET or fellow, the self‐ratings declined. It is important to emphasize that this model is a multiple regression model, so the post‐graduate year has already been factored in. This finding suggests that those in SET and fellows have better insight and are more aware of their own weaknesses in all nine competencies. Trainees early in their career are unconscious of their incompetence and rate themselves higher than they feel they are, that is to say that they ‘don't know what they don't know’. As a trainee then progresses in their training, they then reach a level of conscious incompetence. They start to understand the limits of their knowledge and skills, they ‘know what they don't know’. The last two stages of this model are conscious competence and unconscious competence and are probably most likely to be seen at the consultant and senior consultant level. The instrument used in this study prompts trainees to provide ratings that they feel are relative to their level of experience, this is subjective in nature. However, it does suggest that there is an underlying element of increasing standards and expectations of oneself at higher levels of training. This progression can also be interpreted as improving insight and is likely a result of the actual training program and increasing responsibility in their roles, rather than a simply having more years of experience. For two of the nine competencies ‘health advocacy’ and ‘medical expertise’ the model shows that females and IMGs rate themselves higher than their counterparts. However, the reliability of trainees self‐rating themselves on ‘health advocacy’ is poor so it is difficult to interpret this further and requires further research to clarify this specifically. The finding that females are expected to rate themselves higher than males in ‘medical expertise’ is interesting, as it is not reflected in any of the other competencies. The same finding is also seen for IMGs rating themselves higher than non‐IMGs. There is no literature to date that explains this and further research is required to clarify these findings and investigate the underlying reasons. The pilot of this instrument has its limitations. Firstly, each trainee was only required to nominate two supervisors for rating and feedback. We encourage that further use of this instrument should utilize as many supervisors as possible in as it would help not only provide more volume and variation in feedback, but also identify any underperforming trainees. For an instrument to be reliably used as a 360° assessment, then more colleagues should be sought for assessment as literature suggests as many as 5–10 are required for reproducible results. The competency of ‘health advocacy’ did not have many sub‐questions and can be further expanded to improve reliability. The sample size of Fellows was small and we did not stratify IMG by the country of their original training, nor whether the training was undertaken in an English‐speaking country or not. Further research with more Fellows involved and data around IMGs to find if there are any correlations and allow us to draw stronger conclusions. There would also be value in comparing this new assessment tool against other validated assessment tools in the medical field. This pilot has provided an opportunity to explore the use of a new assessment and feedback instrument for surgical trainees that is aligned to the RACS competency framework using behaviourally anchored questions and rating scales to maximize relevance and trainee reflection upon their skills. Looking ahead, a better understanding of this instrument will potentially be helpful in early identification of underperforming trainees in order to facilitate early intervention, or even its use as a selection tool in formal training programs.

Conflict of interest

None declared.

Author contributions

Tzong‐Yang Pan: Conceptualization; data curation; formal analysis; investigation; methodology; project administration; resources; software; validation; visualization; writing – original draft; writing – review and editing. Frank Piscioneri: Conceptualization; data curation; formal analysis; investigation; methodology; project administration; supervision; validation; writing – review and editing. Cathy Owen: Conceptualization; data curation; formal analysis; investigation; methodology; project administration; resources; supervision; validation; visualization; writing – review and editing. Appendix S1 Supporting information Click here for additional data file.
  5 in total

1.  Self vs expert assessment of technical and non-technical skills in high fidelity simulation.

Authors:  Sonal Arora; Danilo Miskovic; Louise Hull; Krishna Moorthy; Rajesh Aggarwal; Helgi Johannsson; Sanjay Gautama; Roger Kneebone; Nick Sevdalis
Journal:  Am J Surg       Date:  2011-10       Impact factor: 2.565

Review 2.  Multisource feedback to assess surgical practice: a systematic review.

Authors:  Khalid Al Khalifa; Ahmed Al Ansari; Claudio Violato; Tyrone Donnon
Journal:  J Surg Educ       Date:  2013-04-10       Impact factor: 2.891

Review 3.  The reliability, validity, and feasibility of multisource feedback physician assessment: a systematic review.

Authors:  Tyrone Donnon; Ahmed Al Ansari; Samah Al Alawi; Claudio Violato
Journal:  Acad Med       Date:  2014-03       Impact factor: 6.893

4.  Resident-patient interactions: the humanistic qualities of internal medicine residents assessed by patients, attending physicians, program supervisors, and nurses.

Authors:  J O Woolliscroft; J D Howell; B P Patel; D B Swanson
Journal:  Acad Med       Date:  1994-03       Impact factor: 6.893

5.  Validation of Multisource Feedback in Assessing Medical Performance: A Systematic Review.

Authors:  Sebastian Stevens; James Read; Rebecca Baines; Arunangsu Chatterjee; Julian Archer
Journal:  J Contin Educ Health Prof       Date:  2018       Impact factor: 1.355

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.