Literature DB >> 30007931

Evaluating the validity of the selection measures used for the UK's foundation medical training programme: a national cohort study.

Daniel T Smith1, Paul A Tiffin2.   

Abstract

OBJECTIVES: Currently relative performance at medical school (educational performance measure (EPM) decile), additional educational achievements and the score on a situational judgement test (SJT) are used to rank applicants to the UK Foundation Years postgraduate medical training programme. We sought to evaluate whether these three measures were predictive of subsequent successful completion of the programme, and thus were valid selection criteria.
METHODS: Data were obtained from the UK Medical Education Database (UKMED) on 14 131 UK applicants to the foundation programme starting in 2013 and 2014. These data included training outcomes in the form of Annual Reviews of Competency Progression (ARCPs), which indicated whether the programme was successfully completed. The relationship between applicants' performance on the three selection measures to the odds of successful programme completion were modelled.
RESULTS: On univariable analyses, all three measures were associated with the odds of successful completion of the programme. Converting the SJT score to deciles to compare the effect sizes suggested that one decile increase in the EPM increased the odds of completing the programme by approximately 15%, whereas the equivalent value was 8% for the SJT scores. On multivariable analyses (with all three measures included in the model), these effects were only independently and statistically significant for EPM decile (OR 1.14, 95% CI 1.10 to 1.18, p<0.001) and SJT z-score decile (OR 1.05, 95% CI 1.01 to 1.09, p=0.02).
CONCLUSIONS: The EPM decile and SJT scores may be effective selection measures for the foundation programme. However, educational achievements does not add value to the other two measures when predicting programme completion. Thus, its usefulness in this context is less clear. Moreover, our findings suggest that the weighting for the EPM decile score, relative to SJT performance, should be increased. © Author(s) (or their employer(s)) 2018. Re-use permitted under CC BY-NC. No commercial re-use. See rights and permissions. Published by BMJ.

Entities:  

Keywords:  foundation programme; selection

Mesh:

Year:  2018        PMID: 30007931      PMCID: PMC6082493          DOI: 10.1136/bmjopen-2018-021918

Source DB:  PubMed          Journal:  BMJ Open        ISSN: 2044-6055            Impact factor:   2.692


The study covers two complete cohorts of UK applicants to foundation training with the Annual Reviews of Competency Progression (ARCP) outcomes for the first 2 years of their foundation training. Furthermore, there is no ‘range restriction’ (ie, outcomes can only be observed in selected candidates) as nearly all UK applicants get a place on the foundation training programme. There is very little variation captured in ARCP outcomes: most trainees complete the foundation programme. Thus, the measure, in this context, will only yield information on trainees at the lower end of the performance range. More sensitive criteria for assessing the validity of the selection measures would be useful. The study did not have access to more granular data on ‘doctors in difficulty’ captured locally by foundation schools and some of the doctors will have been assisted in order to complete the programme. Similarly, additional information may be present in the end of placements reports completed by trainees’ clinical supervisors in the e-portfolios each trainee is required to maintain.

Introduction

In the UK, the foundation programme is a 2-year generic training programme, which forms the bridge between medical school (4–6 years depending on the school) and specialist medical/general practice training. It is equivalent to the internships in the USA.1 In order to standardise recruitment to this stage of training, the Improving Selection to the Foundation Programme project developed methods for selection into the foundation programme. These were the situational judgement test (SJT) and the educational performance measure (EPM). SJTs present a series of scenarios that depict hypothetical workplace interpersonal dilemmas. In the SJT format used in the foundation programme selection candidates must rank a number of possible behavioural responses, in order of appropriateness and/or effectiveness. A score is generated based on the similarity of the ordering to those previously agreed on by a subject matter expert panel. The content domains of the SJT are labelled as follows: coping with pressure, working effectively as part of a team, effective communication, problem solving and commitment to professionalism.2-4 The SJT scores have been in use for selection to the foundation programme since 2013. Such SJT scores have been shown to have predictive validity for subsequent workplace performance across a range of occupations.5 6 However, they have only been used in medical selection in more recent years and there are fewer validity studies in this context. Previously, a study examining the validity of the SJT for selection into general practice (GP) training reported that the scores accounted for 6% of the variation in end of GP training assessments. However, it is not clear if this finding relates to the applied knowledge test or the clinical skill assessments.7 A separate study examined recruitment to core medical training (CMT).8 The authors concluded that SJT performance was the best single predictor of interview scores. However, the study did not use outcomes from medical training such as Membership of the Royal Colleges of Physicians of the UK exam results to explore the predictive validity of the SJT for selection into CMT programmes. The authors note that the study sample comprised only a subset of applicants—those who applied to both CMT and GP. Thus, this sample may not have been representative of the CMT applicant population as a whole. A follow-up study of this same cohort of CMT trainees reported on the relationship between the selection methods used and subsequent performance at postgraduate membership exams.9 The authors reported that the clinical problem solving test and SJT scores explained an additional 6.3%–21.6% of the variance in performance on the membership exams, after accounting for the variance predicted by the existing CMT selection methods. In relation to the SJT for selection into foundation training, a report on the initial validation study, produced by Work Psychology Group (the company that developed the SJT) has been published.10 11 The authors gathered data on the performance during the first year of the programme (‘F1’) for a sample of 391 F1 doctors across five foundation schools that entered foundation training in 2013 using a bespoke questionnaire, matched to the domains measured at application, completed by supervisors. Their sample specifically targeted doctors in the first year of the foundation programme, who had received particularly high or particularly low SJT scores. The relationship between application scores (SJT and EPM) and the performance scores from the bespoke questionnaire and Annual Reviews of Competency Progression (ARCP) were analysed. They found that higher SJT scores and higher EPM were associated with higher ratings of F1 performance on the questionnaires. They found the correlations were different when their sample was split into high and low SJT scorers. From this analysis, they concluded that the EPM had a stronger correlation with performance for the high scoring SJT group and the SJT a stronger correlation for the low scoring SJT group. This observation could be explained by the contrasting psychometric properties of the two measures: SJTs tend to yield maximal information at the lower range of ability (ie, most candidates find the items relatively easy),12 while EPM is likely to have been able to discriminate between average to high performing candidates (as in effect, the measure is based on rankings with peers in medical school). They found no differences on EPM or SJT scores for those who received unsatisfactory ARCP outcomes compared with those that received satisfactory outcomes. However, they note this was likely to be due to their small sample size. The authors made the following recommendation: "that further studies are undertaken to explore the relationship between performance at application and performance outcomes beyond F1 (for example at the end of F2 and into specialty training) and that application scores (particularly SJT scores) spanning the full range of scores are targeted. If the relationship between application scores and ARCP outcomes is to be examined further, a large population (ideally all schools) should be targeted, as incidences of unsatisfactory ARCP outcomes appear to be very rare (1.1% in the present sample)." The EPM has two elements which this study considers as separate variables: EPM decile score and educational achievements.

The EPM decile score

Students in the graduating cohort are ranked on their medical school performance. Schools were free to decide which assessments to include, provided they met the following criteria: Summative (and hence subject to formal controls); Cover clinical knowledge, skills and performance; Cover non-clinical performance; Cover all aspects of the curriculum assessed up to the end of the penultimate year at medical school; Represent the average performance of the applicants over time, rather than being limited to a snap-shot; Include written and practical forms of assessment. Schools were required to consult with students and publish on their website which assessments they included in the score.3

Educational achievements

These are scored by considering additional degrees (maximum of 5 points, scored 5 points for a PhD through to 1 point for a 2.2 class intercalated degree which does not extend the degree programme) and publications (maximum 2 points, 1 point per publication). There are no published studies on the predictive validity of the EPM decile scores for selection into the foundation years. However, Simon et al reported on the relationship between EPM decile scores and SJT scores, but the data were obtained from trainees in a self-reported survey rather than directly from the UK Foundation Programme Office (UKFPO).13 Their survey achieved a response rate of only 8% (n=3175 from 12 medical schools), so their results (showing no observed association between EPM decile and SJT score) are likely to have been subject to validity-threatening response bias. It is not clear why this approach to the study was taken rather than obtaining the data directly from the UKFPO. These three selection measures are combined into an overall score that is used to rank applicants to the foundation programme, with the EPM decile score and educational achievements combined into the EPM score and given equal weighting to the SJT score.14 Each applicant is allocated in rank order to their highest preference foundation school (a conceptual grouping of medical schools, deaneries and organisations delivering healthcare—trusts or boards depending on the country within the UK) where a place is available. Highly ranked applicants are more likely to be placed in their first preference foundation school. Following entry to the foundation programme, each medical trainee’s progress is reviewed at an ARCP. The rating at ARCP is based on a portfolio of evidence collated in the e-portfolio including reviews from supervisors. ARCPs were introduced into foundation training in 2012.15 This approach to foundation programme selection, and in particular the equal weighting given to the EPM and SJT scores, has attracted criticism. Najim et al noted that an applicant could jeopardise 5–6 years of hard work at medical school through underperformance on a single, 2-hour test.16 With the advent of the UK Medical Education Database (UKMED), it is now possible to link information from the various databases used to administer medical education.17 Thus, there was an opportunity to assess the extent to which the selection scores predicted foundation ARCP outcomes for two cohorts of applicants. This study is thus important in adding to the growing body of evidence in relation to the use of SJTs in medical selection. To the authors’ knowledge, it is the first investigation into whether performance on this measure predicts successful completion of an early postgraduate medical training programme. Moreover, developing a better understanding of how SJT scores should be weighted within the selection process, in relation to academic achievement, is crucial. This need is especially pressing given the recent, rapid and international implementation of SJTs into medical selection over a range of career stages.

Methods

Study population

All applicants making their final application (ie, the application that led to the applicant starting on the foundation programme) in 2013 and 2014 were eligible for inclusion in the analysis (n=15 249). Figure 1 outlines which cases had enough data points for inclusion in the analysis that follows. There were data on 14 131 doctors that met the criteria for inclusion in the main analyses.
Figure 1

Data flow through study. ARCP, Annual Reviews of Competency Progression.

Data flow through study. ARCP, Annual Reviews of Competency Progression. The final sample available for analysis comprised 7134 doctors who started their foundation training in 2013 (50.5%) and 6997 who started in 2014 (49.5%). All doctors had attended a UK medical school, as at this point in its development UKMED did not include graduates from non-UK institutions. The mean age on 1 August in the year the doctor started foundation training was 24.88, with an SD of 2.73 (n=14 131). Their characteristics are further described in table 1.
Table 1

Sample demographics

DemographicGroupN%
SexMan625844.29
Woman787355.71
EthnicityWhite959567.90
Asian or Asian British312222.09
Black or Black British3712.63
Mixed5523.91
Other ethnic groups4323.06
Missing590.42
NationalityOther nationalities163411.56
British12 49788.44
Sample demographics

Patient and public involvement

Neither patients nor the public were involved in this study. Medical students and trainees are represented on the UKMED Advisory Board, which approves UKMED research projects.

Data management

The UKMED includes data from foundation programme applications obtained from the UKFPO’s Application System and MSC Assessment. Educational achievement scores were capped at seven as per the guidance in the UKFPO Applicant Handbook.11 SJT scores are only equated across papers within the application year, so to allow scores from >1 year to be used together, the SJT raw score was converted to a z-score based on the operational statistics for the applicant’s paper and year in the technical reports published by the Work Psychology Group on the Improving Selection to the Foundation Programme website.3 4 To aid interpretation of the ORs in the models used to predict successful completion, we converted this z-score into deciles based on their rank within the year of application to allow direct comparison with the EPM decile. Thus, both predictors were placed on a similar metric, although the locally derived nature of the EPM decile must be borne in mind. The UKMED receives ARCP outcome data from the GMC’s annual collection of ARCP outcomes from postgraduate training providers.18 19 The outcome variable used was obtaining an ‘outcome 6’. This is defined by the UKFPO as ‘satisfactory completion of F2—recommendation for the award of the Foundation Achievement of Competence Document’ at the end of their F2 year in 2015 (2013 starters) or 2016 (2014 starters).14 Medical schools were defined using the values in the Higher Education Statistics Agency (HESA) data. Cases were grouped as follows: The medical school on entry and exit was used when the admitting school and graduating school were the same. Where the graduating medical school was not the same as the admitting medical school and the case was part of a group that formed a distinct cohort with >20 cases, such as students who started at St Andrews, Durham or Oxbridge but completed their clinical undergraduate years at another medical school, we used first and last medical school combined. EPM decile scores are calculated for each medical school’s graduating cohort. We used the foundation school that awarded the outcome 6. In 12 188 (86.3%) cases, the foundation school awarding the outcome 6 was the same as the foundation school awarding the ARCP outcome in the first year of the foundation programme. Some of the 14% of changes reflected renaming and merging of foundation schools, for example, 74 cases started in Staffordshire Foundation School and received their outcome 6 from West Midlands North Foundation School. There were also 906 doctors who started at one of the following foundation schools: North Central, North East Thames or North West Thames Foundation School, but received their outcome 6 from North Thames Foundation Schools. HESA records disability for each academic year the student has a record.20 For these analyses, we used the disability value from the final year. The disability categories recorded by HESA were collapsed into ‘no disability’ versus ‘disability’. Age was calculated as age at the start of the foundation programme: age on 1 August 2013 or 1 August 2014, depending on the year or application.

Analysis

Data management and analysis was conducted using SPSS V.24 and MLwiN V.2.32.21 Multilevel logistic regression models with trainees cross-classified and nested within medical schools and foundation schools were fitted in MLwiN using the method described by Leckie.22 In addition to testing which of the three selection measures predicted completion of the foundation programme, additional analyses were performed to understand if any sociodemographic variables were related to foundation outcomes. This model was fitted using stepwise backwards elimination: at each step, a non-significant variable was eliminated from the model, until only significant variables remained at the p=0.05 level. The stepwise method used a reduced dataset where no missing data on the covariates existed, to ensure ‘true nesting’ while model building (n=7539). The following variables were not statistically significant when entered in the multivariable model: sex, ethnicity, nationality, school type, receipt of a UKCAT bursary, graduate on entry, parental education, IMD-quintile socioeconomic classification, young participation (POLAR3) quintile, Income Deprivation Affecting Children Index rank, course type grouped, year commencing medical school and SJT paper number (see online supplementary material for more information on these variables). Figures 2, 3 and 4 display the mean proportions with 95% CIs. These were calculated using the method recommended by Newcombe and Altman.23 Per cent awarded an outcome 6 by educational performance measure deciles (n=14 131). Per cent awarded an outcome 6 by situational judgement test z-score deciles (n=14 128). Per cent awarded an outcome 6 by educational achievements score (n=14 131).

Results

A total of 13 788 (97.6%) cases in the sample had successfully achieved the required competencies (ARCP ‘outcome 6’) by the end of their 2-year programme. The results from univariable logistic regressions (table 2) show that all three measures, the SJT z-score (mean=0.10, SD=0.86), EPM (mean=5.62, SD=2.83) and educational achievements (mean=2.47, SD=2.01) predict obtaining an outcome 6. For each decile increase in the EPM, an applicant achieves the odds of an ‘outcome 6’ increase by roughly 15%; for each decile, increase in the SJT decile score an applicant achieves the odds of an outcome 6 increase by approximately 8%. For each additional point in their educational achievement score, a foundation applicant’s odds of an outcome 6 increase by roughly 7%.
Table 2

Univariable relationships between each foundation selection measure and successful completion (ARCP ‘outcome 6’)

Predictor variable (n)OR95% CI for ORP values
EPM decile score (n=14 131)1.1521.107 to 1.198<0.001
SJT z-score decile (n=14 128)1.0761.036 to 1.118<0.001
SJT z-score (n=14 128)1.2941.152 to 1.452<0.001
SJT equated score (n=14 131)1.0081.005 to 1.012<0.001
Educational achievements (n=14 131)1.0651.010 to 1.1240.021
Univariable relationships between each foundation selection measure and successful completion (ARCP ‘outcome 6’) Figures 2 and 3 illustrate the relationship between completion of the foundation programme and the SJT z-score and EPM deciles: for both measures those applicants achieving at the lowest decile have a reduced proportion of ‘outcomes 6’ compared to those at the top decile, but the relationship for the intermediate deciles is non-linear. The distribution of the educational achievements score does not allow splitting into decile groups, so the relationship between the proportion of ‘outcome 6 s’ and the score is presented separately on figure 4. The groups with scores of 0, 2, 3 and 7 have the same mean proportion of outcome 6.
Figure 4

Per cent awarded an outcome 6 by educational achievements score (n=14 131).

The three selection scores intercorrelated to some degree: the highest correlation being between EPM decile and SJT z-score (Spearman’s rho=0.30, N=14 128) and the lowest between educational achievements and SJT z-score (Spearman’s rho=0.12, n=14 131), with the correlation between EPM decile and educational achievements at rho=0.28 (n=14 131). In the first model (model 1, table 3), a two-level multivariable logistic regression was fitted to predict achievement of an ‘outcome 6’ with trainees nested within a cross-classification of medical schools and foundation schools. The EPM decile score and the SJT z-score were statistically significant predictors at the p<0.05 level. In terms of interpreting the results, for an increase of one EPM the odds of successful completion of the foundation programme increased by approximately 14%. Similarly, for an increase of 1 SD on the SJT z-score the odds of an outcome 6 increased by approximately 18%. The education achievement measure is not independently and statistically significantly associated with successful completion when all three measures are included in a model.
Table 3

Results from multilevel multivariable logistic regression analyses predicting an ‘outcome 6’ with the foundation selection measures (N=14 128 applicants nested in medical schools and foundation schools (cross-classified))

VariableOR95% CI for ORP values
Model 1EPM decile score1.1401.095 to 1.185<0.001
SJT z-score1.1831.008 to 1.2590.011
Educational achievements0.9980.937 to 1.0590.944
Model 2EPM decile score1.1411.098 to 1.184<0.001
SJT z-score decile1.0491.010 to 1.0880.017
Educational achievements0.9950.932 to 1.0580.867
Model 3EPM decile score1.1131.072 to 1.154<0.000
SJT z-score1.1220.996 to 1.2470.075
Educational achievements1.0610.998 to 1.1230.068
No disability1.6601.345 to 1.9760.002
Number of applications0.5440.154 to 0.9340.002
Age0.9040.873 to 0.935<0.000
Results from multilevel multivariable logistic regression analyses predicting an ‘outcome 6’ with the foundation selection measures (N=14 128 applicants nested in medical schools and foundation schools (cross-classified)) In the second model (model 2, table 3), the SJT z-score was converted to the same metric as the EPM (decile ranks within year of application to foundation) to allow the adjusted ORs obtained for the two measures to be directly compared. As before, one decile increase in the EPM increases the odds of an outcome 6 by 14%; one decile increase in the rank of the SJT z-score independently increases the chance of an outcome 6 by 5% (p=0.02). The variables that were statistically significant predictors at the p<0.05 level in a series of univariable analyses (see online supplementary material) were included in a third multivariable model (model 3, table 3). As there were no missing data for the final variables included, it was possible to use all cases. Trainees with no disability recorded by HESA in their final year of medical school had an approximately 66% higher odds of an outcome 6 compared with those with a disability recorded. Older trainees were less likely to be awarded an outcome 6: each additional year reduced the odds by approximately 9%. Trainees who had made additional applications were less likely to get an outcome 6, with each additional application reducing the odds by roughly 54%. After controlling for age, disability, the number of applications made and the medical school, the EPM decile score remained a significant predictor of obtaining an outcome 6 while SJT z-score did not.

Discussion

In this study, we were able to show that the three main selection measures (SJT score, EPM and educational achievements) used in ranking applicants to the UK foundation programme were related to the odds of successful completion of this stage of training. However, the relationship between educational achievements and completion was not independent of the other two measures. In contrast, the SJT score appeared to offer some degree of incremental predictive validity over that provided by the EPM deciles, suggesting that it is capturing additional, and relevant, information on applicants as intended by its developers. As such, our findings add to the emerging evidence of the potential usefulness and validity of SJTs in medical selection across different career stages and for different clinical specialisations in relation to a number of educationally and clinically relevant outcomes. 24 25 Our findings in relation to trainee age are in keeping with those reported by an earlier study by Pyne and Ben-Shlomo,26 who reported that older doctors in their sample of specialty trainees were more likely to have problems with progression at ARCP than their younger colleagues. The relationship between disability and ARCP has not previously been reported on; here, we found that those trainees who had a disability (any category) as recorded by HESA were less likely to obtain an outcome 6. Reflecting on our key findings, it is perhaps unsurprising that the educational achievements measure was not independent of the other measures. This metric would have been skewed by the relatively small number of students who had participated in research projects, often as part of intercalated degrees. Such students would also have been likely to have been ranked relatively highly, according to their medical school EPMs. It is also worth commenting that, at first glance, both the EPM deciles and the SJT scores appear equally predictive of completion of the foundation programme. However, when we attempt to place both measures, although be crudely, on the same scale (ie, divided into deciles), it is clear that EPM deciles are more predictive of this outcome, compared with SJT scores, with a ratio of roughly 2:1. Of course, the situation is complicated by the fact that EPM deciles are locally derived measures, while SJTs are nationally standardised tests. There may be ways in future research that the EPMs can be adjusted to make them more nationally comparable using ‘Peer-Competition Rescaling’.27 Nevertheless, despite the local nature of EPM deciles this finding remains relevant to policy as both SJT scores and EPM deciles are used in the national selection process. It is also known that SJTs are generally encountered as relatively easy tests by candidates, and therefore most of the information is available on those below the average level of performance.12 This infers that the SJT scores are likely to be relatively poor at differentiating more highly performing candidates from each other. We noted that once the scores were divided into deciles, this would have inevitably led to some loss of information, and resulted in a reduced degree of statistical significance in the relationship between this predictor and the outcome of interest, with the p value increasing from 0.01 to 0.02 when entered into a multivariable model with the other selection measures. In contrast, EPM deciles may be able to differentiate between both low and high performing candidates. This proposition is supported by our findings illustrated in figure 2: the centre portion of the graph is relatively flat with the slopes being steepest at the extreme ends of lower and highly EPM decile ranked applicants. We also noted that, in contrast to EPM, SJT scores were not independently and statistically significant predictors of outcome 6 s once three other background variables were put into a multivariable model (disability, number of applications and age). However, as applicant ranking does not take into account these latter three factors, this finding does not have direct importance to policy. Furthermore, it does suggest that at least some of the variance in SJT scores may be associated with these variables, although the directionality of influence could not be established from these data. Nevertheless, it is known that some demographic factors (such as female sex and ethnicity) are associated with SJT performance in general28 and therefore the use of such selection measures may have an indirect effect on the advantage (or disadvantage) experienced by certain subpopulations of applicant.
Figure 2

Per cent awarded an outcome 6 by educational performance measure deciles (n=14 131).

This was a relatively complete national dataset with few missing data. However, a number of limitations of the study are worth noting. First, the UKMED phase I cohort does not include graduates from non-UK medical schools who apply to foundation training. However, from 2015 onwards the non-UK students have been included in the data UKMED receives on foundation applications, permitting research into this area. Second, suboptimal ARCP outcomes are relatively rare in the foundation programme period of training. Unlike later periods of postgraduate medical training, there are fewer categories of ARCP outcome, and therefore at foundation stage they are a relatively information poor variable, picking up only cases where issues impacting performance were severe enough to prevent the doctor completing the programme. Moreover, there were no data on which trainees received remedial support during the foundation programme were available for this study. Thus, it is possible that a number of doctors who received outcome 6 s had required, sometimes considerable, support in order to achieve this. To establish whether the SJT and EPM deciles predict which trainees required additional support from their foundation schools would require the UKMED to have identifiable data on doctors in difficulty on the foundation programme. The UKFPO annual reports give figures for the numbers of doctors in difficulty: in 2014 there were 186 F1s and 163 F2s from UKMED medical school monitored via foundation schools’ doctors in difficulty (DiD) policies and processes. In 2015, there were 251 F1s and 210 F2s monitored via these policies.29 We do not know which ARCP outcomes these doctors were finally awarded. Similarly, at the other end of the scale, the outcome 6 category does not capture performance that exceeds that which is required to achieve the foundation competencies. Thus, in psychometric terms, the outcome would have generally yielded information on trainees at the lower end of performance. That is, the results may tell us relatively little about candidates who score relatively highly on the three selection measures. More variance may also be present in the end of placement reports completed by trainees’ clinical supervisors in the e-portfolio each trainee is required to maintain. A score derived by summing across the six to eight clinical supervisor reports that are completed over the course of the foundation programme may provide a more granular outcome measure. Such an approach may be considered closer to the criterion used by Patterson et al, who employed supervisor ratings of trainee job performance at 1 year into training as an outcome to validate selection measures for entry to GP training against.7 Some variation in ARCP outcomes across foundation schools was noted. Moreover, EPM is a local rather than national measure. However, the use of a cross-classified multilevel model should have controlled for these potential clustering effects on the outcome of interest. The present study relied on data that were already routinely collected and not specifically designed as a criterion to assess the validity of the selection measures. As Austin and Villanova noted using a general overall performance construct such as successful completion of the programme may not allow adequate matching to the predictive measures and it may be more fruitful to match selection measures to particular criteria representing particular aspects of job performance.30 Nevertheless, trainees undertaking the foundation programme and those responsible for managing the programmes are likely to regard it as an important criterion even if fails to capture the full range of foundation doctors’ performance. Other possible criteria include the multispecialty recruitment assessments used for selection into the next stage of training including GP training programmes.31 Our findings have clear indications for selection policy into the foundation programme. In particular, they provide evidence to support the intuitive sense, expressed by some, that excessive weight may be given to the SJT scores within the ranking process.16 Rather our results would suggest a 2:1 weighting ratio between the EPM decile and the SJT score may be the optimum, if the aim is to rank and advantage candidates most likely to successfully complete the foundation programme. Policywise, our findings provide less evidence to support using the educational achievements score in calculating applicant rankings for the foundation programme. If the aim is to select candidates most likely to complete this phase of training then our results suggest that the educational achievements score does not add any incremental value beyond the EPM deciles or SJT scores. However, we are aware we have used a somewhat circumscribed outcome, that is, completion of the programme. It may be that the selectors wish to also advantage applicants, according to their academic records, which are most likely to be educationally, and perhaps clinically successful in their long-term careers. If removal of the educational achievements and an increase in the weighting of the EPM deciles were to be considered then an exercise could be undertaken to model whether such a change would have a significant impact on the numbers of applicants whose rank changes to an extent that they would have been offered a different unit of application. Such weightings would be more in-line with research in selection outside of medicine, where measures of ability are more strongly predictive than personality-type measures. Schmidt and Hunter reviewed meta-analyses of selection methods: when predicting performance in job training programmes they report a mean correlation coefficient of 0.56 for cognitive ability, whereas for integrity and conscientiousness tests they report correlation coefficients of 0.38 and 0.30.32 One might argue that the EPM measure will inevitably reflect cognitive ability while the SJT is more akin to an integrity or personality test. Further research could focus on the extent to which these foundation selection measures predict long-term success and career choices in medicine. Moreover, there is an immediate intention to further explore and describe the types of disability reported by medical graduates in a forthcoming UKMED project: UKMEDP54 declared disability in the UKMED dataset 2002–2016: an exploratory descriptive analysis.33 In conclusion, the continued use of the SJT in selection into the foundation programme is justified by these findings, although it may be that excessive weight is being placed on the score, relative to the other two measures. Moreover, the UKFPO could consider discontinuing the use of educational achievements for ranking applicants to the foundation programme, unless it is shown to be associated with more distal markers of success in medical training.
  11 in total

Review 1.  Situational judgement tests in medical education and training: Research, theory and practice: AMEE Guide No. 100.

Authors:  Fiona Patterson; Lara Zibarras; Vicki Ashworth
Journal:  Med Teach       Date:  2015-08-27       Impact factor: 3.650

2.  New machine-marked tests for selection into core medical training: evidence from two validation studies.

Authors:  Fiona Patterson; Victoria Carr; Lara Zibarras; Bill Burr; Liz Berkin; Simon Plint; Bill Irish; Simon Gregory
Journal:  Clin Med (Lond)       Date:  2009-10       Impact factor: 2.659

3.  Does a high ranking mean success in the Situational Judgement Test?

Authors:  Elizabeth Simon; Keith Walsh; Flora Paterson-Brown; David Cahill
Journal:  Clin Teach       Date:  2015-02

4.  The Predictive Validity of a Text-Based Situational Judgment Test in Undergraduate Medical and Dental School Admissions.

Authors:  Fiona Patterson; Fran Cousans; Helena Edwards; Anna Rosselli; Sandra Nicholson; Barry Wright
Journal:  Acad Med       Date:  2017-09       Impact factor: 6.893

5.  Evaluating the complementary roles of an SJT and academic assessment for entry into clinical practice.

Authors:  Fran Cousans; Fiona Patterson; Helena Edwards; Kim Walker; John C McLachlan; David Good
Journal:  Adv Health Sci Educ Theory Pract       Date:  2017-02-08       Impact factor: 3.853

6.  The predictive validity of a situational judgement test, a clinical problem solving test and the core medical training selection methods for performance in specialty training .

Authors:  Fiona Patterson; Safiatu Lopes; Stephen Harding; Emma Vaux; Liz Berkin; David Black
Journal:  Clin Med (Lond)       Date:  2017-02       Impact factor: 2.659

7.  Older doctors and progression through specialty training in the UK: a cohort analysis of General Medical Council data.

Authors:  Yvette Pyne; Yoav Ben-Shlomo
Journal:  BMJ Open       Date:  2015-02-03       Impact factor: 2.692

8.  The situational judgement test: a student's worst nightmare.

Authors:  Muhammad Najim; Riham Rabee; Yusuf Sherwani; Maroof Ahmed; Muhammad Ashraf; Osama Al-Jibury; Rula Najim
Journal:  Adv Med Educ Pract       Date:  2015-10-05

9.  The UK medical education database (UKMED) what is it? Why and how might you use it?

Authors:  Jon Dowell; Jennifer Cleland; Siobhan Fitzpatrick; Chris McManus; Sandra Nicholson; Thomas Oppé; Katie Petty-Saphon; Olga Sierocinska King; Daniel Smith; Steve Thornton; Kirsty White
Journal:  BMC Med Educ       Date:  2018-01-05       Impact factor: 2.463

10.  The predictive validity of a situational judgement test and multiple-mini interview for entry into postgraduate training in Australia.

Authors:  Fiona Patterson; Emma Rowett; Robert Hale; Marcia Grant; Chris Roberts; Fran Cousans; Stuart Martin
Journal:  BMC Med Educ       Date:  2016-03-09       Impact factor: 2.463

View more
  13 in total

1.  Exploring UK medical school differences: the MedDifs study of selection, teaching, student and F1 perceptions, postgraduate outcomes and fitness to practise.

Authors:  I C McManus; Andrew Christopher Harborne; Hugo Layard Horsfall; Tobin Joseph; Daniel T Smith; Tess Marshall-Andon; Ryan Samuels; Joshua William Kearsley; Nadine Abbas; Hassan Baig; Joseph Beecham; Natasha Benons; Charlie Caird; Ryan Clark; Thomas Cope; James Coultas; Luke Debenham; Sarah Douglas; Jack Eldridge; Thomas Hughes-Gooding; Agnieszka Jakubowska; Oliver Jones; Eve Lancaster; Calum MacMillan; Ross McAllister; Wassim Merzougui; Ben Phillips; Simon Phillips; Omar Risk; Adam Sage; Aisha Sooltangos; Robert Spencer; Roxanne Tajbakhsh; Oluseyi Adesalu; Ivan Aganin; Ammar Ahmed; Katherine Aiken; Alimatu-Sadia Akeredolu; Ibrahim Alam; Aamna Ali; Richard Anderson; Jia Jun Ang; Fady Sameh Anis; Sonam Aojula; Catherine Arthur; Alena Ashby; Ahmed Ashraf; Emma Aspinall; Mark Awad; Abdul-Muiz Azri Yahaya; Shreya Badhrinarayanan; Soham Bandyopadhyay; Sam Barnes; Daisy Bassey-Duke; Charlotte Boreham; Rebecca Braine; Joseph Brandreth; Zoe Carrington; Zoe Cashin; Shaunak Chatterjee; Mehar Chawla; Chung Shen Chean; Chris Clements; Richard Clough; Jessica Coulthurst; Liam Curry; Vinnie Christine Daniels; Simon Davies; Rebecca Davis; Hanelie De Waal; Nasreen Desai; Hannah Douglas; James Druce; Lady-Namera Ejamike; Meron Esere; Alex Eyre; Ibrahim Talal Fazmin; Sophia Fitzgerald-Smith; Verity Ford; Sarah Freeston; Katherine Garnett; Whitney General; Helen Gilbert; Zein Gowie; Ciaran Grafton-Clarke; Keshni Gudka; Leher Gumber; Rishi Gupta; Chris Harlow; Amy Harrington; Adele Heaney; Wing Hang Serene Ho; Lucy Holloway; Christina Hood; Eleanor Houghton; Saba Houshangi; Emma Howard; Benjamin Human; Harriet Hunter; Ifrah Hussain; Sami Hussain; Richard Thomas Jackson-Taylor; Bronwen Jacob-Ramsdale; Ryan Janjuha; Saleh Jawad; Muzzamil Jelani; David Johnston; Mike Jones; Sadhana Kalidindi; Savraj Kalsi; Asanish Kalyanasundaram; Anna Kane; Sahaj Kaur; Othman Khaled Al-Othman; Qaisar Khan; Sajan Khullar; Priscilla Kirkland; Hannah Lawrence-Smith; Charlotte Leeson; Julius Elisabeth Richard Lenaerts; Kerry Long; Simon Lubbock; Jamie Mac Donald Burrell; Rachel Maguire; Praveen Mahendran; Saad Majeed; Prabhjot Singh Malhotra; Vinay Mandagere; Angelos Mantelakis; Sophie McGovern; Anjola Mosuro; Adam Moxley; Sophie Mustoe; Sam Myers; Kiran Nadeem; Reza Nasseri; Tom Newman; Richard Nzewi; Rosalie Ogborne; Joyce Omatseye; Sophie Paddock; James Parkin; Mohit Patel; Sohini Pawar; Stuart Pearce; Samuel Penrice; Julian Purdy; Raisa Ramjan; Ratan Randhawa; Usman Rasul; Elliot Raymond-Taggert; Rebecca Razey; Carmel Razzaghi; Eimear Reel; Elliot John Revell; Joanna Rigbye; Oloruntobi Rotimi; Abdelrahman Said; Emma Sanders; Pranoy Sangal; Nora Sangvik Grandal; Aadam Shah; Rahul Atul Shah; Oliver Shotton; Daniel Sims; Katie Smart; Martha Amy Smith; Nick Smith; Aninditya Salma Sopian; Matthew South; Jessica Speller; Tom J Syer; Ngan Hong Ta; Daniel Tadross; Benjamin Thompson; Jess Trevett; Matthew Tyler; Roshan Ullah; Mrudula Utukuri; Shree Vadera; Harriet Van Den Tooren; Sara Venturini; Aradhya Vijayakumar; Melanie Vine; Zoe Wellbelove; Liora Wittner; Geoffrey Hong Kiat Yong; Farris Ziyada; Oliver Patrick Devine
Journal:  BMC Med       Date:  2020-05-14       Impact factor: 8.775

2.  Relationship between sociodemographic factors and specialty destination of UK trainee doctors: a national cohort study.

Authors:  Ben Kumwenda; Jennifer Cleland; Gordon Prescott; Kim Walker; Peter Johnston
Journal:  BMJ Open       Date:  2019-03-27       Impact factor: 2.692

3.  National recruitment scheme for pre-registration pharmacist training in England and Wales: a mixed method evaluation of experiences of applicant pharmacy students.

Authors:  Laura McEwen-Smith; Malcolm James Price; Gail Fleming; Tim Swanwick; Christine Hirsch; Asma Yahyouche; Jonathan Ward; Sharon Buckley; Atif Shamim; Vibhu Paudyal
Journal:  BMC Med Educ       Date:  2019-12-04       Impact factor: 2.463

4.  Are there differences between those doctors who apply for a training post in Foundation Year 2 and those who take time out of the training pathway? A UK multicohort study.

Authors:  Jennifer Cleland; Gordon Prescott; Kim Walker; Peter Johnston; Ben Kumwenda
Journal:  BMJ Open       Date:  2019-11-24       Impact factor: 2.692

5.  Virtual Interview, Real Anxiety: Prospective Evaluation of a Focused Teaching Programme on Confidence Levels Among Medical Students Applying for Academic Clinical Posts.

Authors:  Agata P Zielinska; Jamie A Mawhinney; Natalie Grundmann; Sosipatros Bratsos; Jamie Sin Ying Ho; Ankur Khajuria
Journal:  Adv Med Educ Pract       Date:  2021-06-17

6.  Dental foundation training in the COVID-19 era - the national recruitment lottery.

Authors:  Christopher C Donnell; Benjamin M Walker; Jennifer I Foley
Journal:  Br Dent J       Date:  2021-07-09       Impact factor: 1.626

7.  A comparison of undergraduate outcomes for students from gateway courses and standard entry medicine courses.

Authors:  Sally Curtis; Daniel Smith
Journal:  BMC Med Educ       Date:  2020-01-03       Impact factor: 2.463

8.  Does 'online confidence' predict application success and later academic performance in medical school? A UK-based national cohort study.

Authors:  Paul A Tiffin; Lewis W Paton
Journal:  BMJ Open       Date:  2019-12-29       Impact factor: 2.692

9.  '10% of your medical students will cause 90% of your problems': a prospective correlational study.

Authors:  Marina Sawdon; J C McLachlan
Journal:  BMJ Open       Date:  2020-11-04       Impact factor: 2.692

10.  Predictors of ophthalmology career success (POCS) study.

Authors:  Aditi Das; Daniel Smith; Rashmi G Mathew
Journal:  BMJ Open Ophthalmol       Date:  2021-07-12
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.