| Literature DB >> 24645665 |
Johanna C Hissbach, Susanne Sehner, Sigrid Harendza, Wolfgang Hampe1.
Abstract
BACKGROUND: Multiple mini-interviews (MMIs) are a valuable tool in medical school selection due to their broad acceptance and promising psychometric properties. With respect to the high expenses associated with this procedure, the discussion about its feasibility should be extended to cost-effectiveness issues.Entities:
Mesh:
Year: 2014 PMID: 24645665 PMCID: PMC3995077 DOI: 10.1186/1472-6920-14-54
Source DB: PubMed Journal: BMC Med Educ ISSN: 1472-6920 Impact factor: 2.463
Changes made to the procedure (2009 – 2010)
| Number of target variables | 8 | 3 |
| Number of stations | 12 | 9 |
| Duration of interview cycle (minutes) | 78 | 58.5 |
| Number of rounds | 4 | 5 |
| Number of circuits | 2 | 4 |
| Average ratings per candidate and station | 2.3 | 2.17 |
| Number of ratings per candidate | 28 | 19.5 |
| Number of candidates | 80 | 200 |
| Rater training hours | 2 | 4 |
| Interviewers give ratings | Yes | No |
| Score sheets | No anchoring | Anchoring, personalization |
| Check-in time on interview day | Simultaneous | Variable |
Descriptive statistics of candidates and raters
| | ||||
|---|---|---|---|---|
| Raters | 57 | | 94 | |
| Females | 23 | (40%) | 50 | (53%) |
| Candidates invited | 80 | | 200 | |
| Candidates tested | 78 | (98%) | 193 | (97%) |
| Females | 50 | (64%) | 112 | (58%) |
| German | 71 | (91%) | 185 | (96%) |
| EU | 2 | (3%) | 5 | (2%) |
| Non-EU | 5 | (6%) | 3 | (2%) |
| | ||||
| Age | 20.2 | (4.1) | 19.8 | (1.6) |
| GPA (highschool) | 1.7 | (0.2) | 1.7 | (0.2) |
| % achieved of max. HAM-Int score | 68.7 | | 64.7 | |
| Mean station result | 3.43a | (0.83) | 3.23b | (1.05) |
ascale from 0–5; bscale from 1–5.
Station characteristics
| | |||||||||
|---|---|---|---|---|---|---|---|---|---|
| 1 | Personal dilemma I (SP) | 3.97 | (0.97) | | | 0.51 | | 0.42 | |
| 2 | Empathetic communication I (SP) | 3.33 | (1.04) | | | 0.42 | | 0.75 | |
| 3 | Conflict resolution I (SP) | 3.37 | (1.08) | | | 0.36 | | 0.78 | |
| 4 | Motivation interview I | 3.50 | (1.09) | | | 0.48 | | 0.58 | |
| 5 | Ethical dilemma I | 3.17 | (0.93) | | | 0.37 | | 0.65 | |
| 6 | Evaluation social interaction | 3.51 | (1.01) | | | 0.40 | | 0.65 | |
| 7 | Conflict resolution II (SP) | 3.27 | (0.99) | | | 0.10 | | 0.44 | |
| 8 | Decision making | 3.72 | (0.85) | | | 0.30 | | 0.50 | |
| 9 | Conflict resolution II | 3.29 | (0.86) | 3.36 | (1.02) | 0.42 | 0.35 | 0.77 | 0.71 |
| 10 | Breaking bad news (SP) | 3.80 | (0.89) | 2.92 | (1.07) | 0.61 | 0.40 | 0.70 | 0.84 |
| 11 | Clarification of a situation (SP) | 3.18 | (1.04) | 3.09 | (1.13) | 0.33 | 0.38 | 0.73 | 0.57 |
| 12 | Ethical dilemma II | 3.01 | (1.21) | 3.22 | (0.98) | 0.31 | 0.27 | 0.74 | 0.78 |
| 13 | Conflict resolution III (SP) | | | 3.36 | (1.02) | | 0.45 | | 0.55 |
| 14 | Empathetic communication II (SP) | | | 3.19 | (0.98) | | 0.36 | | 0.83 |
| 15 | Decision making II | | | 2.96 | (0.97) | | 0.37 | | 0.78 |
| 16 | Conflict resolution III (SP) | | | 3.29 | (1.24) | | 0.41 | | 0.64 |
| 17 | Motivation interview II | | | 3.68 | (1.02) | | 0.22 | | 0.61 |
| | Mean all stations | 3.43 | (0.83) | 3.23 | (1.05) | 0.38 | 0.36 | | |
| | All two-rater stations | | | | | | | 0.68 | 0.71 |
| All three-rater stations | 0.67 | 0.80 | |||||||
ascale from 0–5; bscale from 1–5.
Variance components of model 1 (without adjustment for rater stringency)
| | |||||
|---|---|---|---|---|---|
| Cand | Systematic differences in candidate performance | 0.13 (0.09-0.20) | 12.2 | 0.21 (0.16-0.27) | 18.1 |
| Stat | Systematic differences in station difficulty | 0.06 (0.02-0.19) | 5.0 | 0.04 (0.01-0.13) | 3.3 |
| Rater | Systematic differences in rater stringency | 0.09 (0.05-0.14) | 7.8 | 0.04 (0.03-0.07) | 3.8 |
| Cand*stat | Differences in candidate performance between stations | 0.30 (0.25-0.36) | 27.6 | 0.39 (0.34-0.43) | 33.3 |
| Rater*cand (+ error) | Differences in rater candidate taste, and residual variance | 0.52 (0.48-0.56) | 47.5 | 0.48 (0.45-0.51) | 41.5 |
| Total | 1.09 | 100 | 1.16 | 100 | |
Rel = Vcand./(Vcand + Vstat/nstat + Vrater/nratings per cand + Vcand*stat/nstat + Vrater*cand/nratings per cand).
2009: Rel = Vcand/(Vcand + Vstat/12 + Vrater/28 + Vcand*stat/12 + Vrater*cand/28) = 0.72.
2010: Rel = Vcand/(Vcand + Vstat/9 + Vrater/19.5 + Vcand*stat/9 + Vrater*cand/19.5) = 0.74.
Variance components of model 2 (reliability model with adjustment for rater stringency)
| | |||||
|---|---|---|---|---|---|
| Cand | Systematic differences in candidate performance | 0.13 (0.08-0.20) | 13.9 | 0.21 (0.16-0.28) | 19.6 |
| Cand*stat | Differences in candidate performance between stations | 0.30 (0.02-0.19) | 31.6 | 0.38 (0.34-0.43) | 35.6 |
| Rater*cand (+ error) | Adjusted differences in rater candidate taste, and residual variance | 0.52 (0.55-0.64) | 54.5 | 0.48 (0.45-0.51) | 44.8 |
| Total | 0.95 | 100 | 1.07 | 100 | |
2009: Rel = Vcand/(Vcand + Vcand*stat/12 + Vrater*cand/28) = 0.75.
2010: Rel = Vcand/(Vcand + Vcand*station/9 + Vrater*cand/19.5) = 0.76.
Figure 1Estimation of total test reliability with different numbers of raters and stations (2009 vs. 2010).
Costs of the 2009 procedure
| Writing and adaptation of 12 scenarios and score sheets | 475 | 23,750 | | | | | | | | |
| Organization of test runs | | | 24 | 720 | | | | | | |
| Test runs | 180 | 9,000 | 24 | 720 | 36 | 360 | 36 | 468 | 1,500 | |
| Room acquisition | | | 3 | 90 | | | | | | |
| Rater recruitment and allocation | 3 | 150 | 15 | 450 | | | | | | |
| Training of actors | 50 | 2,500 | | | 150 | 1,500 | | | | |
| Invitation of / feedback to candidates | | | 5 | 150 | | | | | | |
| Printing of score sheets | | | 5 | 150 | | | 10 | 130 | 200 | |
| Preparation and clearance of rooms | 5 | 250 | 10 | 300 | | | 25 | 325 | | |
| Rater training and testing of candidates | 480 | 24,000 | 25 | 750 | 150 | 1,500 | 120 | 1,560 | | |
| Catering for raters | | | | | | | | | 1,500 | |
| Scanning of score sheets and data analysis | 10 | 500 | 20 | 600 | | | | | | |
| Additional costs for first implementation | 100 | 5,000 | 100 | 3,000 | | | 50 | 650 | 1,500 | |
Costs of the 2010 procedure
| | |||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Adaptation of 4 scenarios and score sheets | 75 | 3,750 | | | | | | | | | | | |
| Writing of 5 new scenarios and score sheets | 200 | 10,000 | | | | | | | | | | | |
| Organization of test runs | | | 4 | 120 | | | | | | | | | |
| Test runs | 80 | 4,000 | 8 | 240 | 10 | 100 | 25 | 325 | 400 | | | | |
| 2,104a | | | |||||||||||
| Room acquisition | | | 10 | 300 | | | | | | | | 300 | |
| Rater recruitment and allocation | 3 | 150 | 20 | 600 | | | | | | | | 750 | |
| Training of actors | 30 | 1,500 | | | 150 | 1,500 | | | | 222a | | 1,000 | |
| Invitation of / feedback to candidates | | | 15 | 450 | | | | | | | | 450 | |
| Printing of score sheets | | | 10 | 300 | | | 20 | 260 | 700 | | | 1,260 | |
| Preparation and clearance of rooms | 5 | 250 | 10 | 300 | | | 30 | 390 | | | | 940 | |
| Rater training and testing of candidates | 995 | 49,750 | 50 | 1,500 | 400 | 4,000 | 200 | 2,600 | | | 2,551b | 8,100 | |
| Interviewer training and testing of candidates | 205 | 10,250 | | | | | | | | 1,139a | | | |
| Catering for raters | | | | | | | | | 4000 | | 170b | 680 | |
| Scanning of score sheets and data analysis | 10 | 500 | 20 | 600 | | | | | | | | 1,100 | |
| | | | |||||||||||
| 3,465 | 2,721 | 14,581 | |||||||||||
aaverage costs per station (costs divided by the number of stations (9)); baverage costs per rater (costs divided by the number of raters per candidate (19.5)).
Figure 2Interrelation of costs and reliability (2010 HAM-Int procedure).