Literature DB >> 33447662

Comparing Standard Setting Methods for Objective Structured Clinical Examinations in a Caribbean Medical School.

Neelam Rekha Dwivedi1, Narasimha Prasad Vijayashankar1, Manisha Hansda1, Arun Kumar Dubey1, Fidelis Nwachukwu1, Vernon Curran2, Joseph Jillwin1.   

Abstract

BACKGROUND: OSCE are widely used for assessing clinical skills training in medical schools. Use of traditional pass fail cut off yields wide variations in the results of different cohorts of students. This has led to a growing emphasis on the application of standard setting procedures in OSCEs. PURPOSE/AIM: The purpose of the study was comparing the utility, feasibility and appropriateness of 4 different standard setting methods with OSCEs at XUSOM.
METHODS: A 15-station OSCE was administered to 173 students over 6 months. Five stations were conducted for each organ system (Respiratory, Gastrointestinal and Cardiovascular). Students were assessed for their clinical skills in 15 stations. Four different standard setting methods were applied and compared with a control (Traditional method) to establish cut off scores for pass/fail decisions.
RESULTS: OSCE checklist scores revealed a Cronbach's alpha of 0.711, demonstrating acceptable level of internal consistency. About 13 of 15 OSCE stations performed well with "Alpha if deleted values" lower that 0.711 emphasizing the reliability of OSCE stations. The traditional standard setting method (cut off score of 70) resulted in highest failure rate. The Modified Angoff Method and Relative methods yielded the lowest failure rates, which were typically less than 10% for each system. Failure rates for the Borderline methods ranged from 28% to 57% across systems.
CONCLUSIONS: In our study, Modified Angoff method and Borderline regression method have shown to be consistently reliable and practically suitable to provide acceptable cut-off score across different organ system. Therefore, an average of Modified Angoff Method and Borderline Regression Method appeared to provide an acceptable cutoff score in OSCE. Further studies, in high-stake clinical examinations, utilizing larger number of judges and OSCE stations are recommended to reinforce the validity of combining multiple methods for standard setting.
© The Author(s) 2020.

Entities:  

Keywords:  OSCE; and modified Angoff’s method; borderline regression method; mean borderline group method; relative method; standard setting; traditional method

Year:  2020        PMID: 33447662      PMCID: PMC7780167          DOI: 10.1177/2382120520981992

Source DB:  PubMed          Journal:  J Med Educ Curric Dev        ISSN: 2382-1205


Introduction

Xavier University School of Medicine (XUSOM) is an offshore medical school located on the island of Aruba. It offers a 4-year MD program similar to North American medical schools. Year 1 and 2 curriculum consist of basic sciences portion which is organized as an integrated organ- based system with both horizontal and vertical integration. Basic science is taught in a hybrid curriculum, using a combination of didactic lectures, Problem Based Learning (PBL), Team Based Learning (TBL), Clinical case presentations and other self-directed learning (SDL) teaching and learning methodologies. The clinical skills training begins in the first semester of medical school as part of an “Early Clinical Exposure” course, where students visit local family physicians’ and specialists’ clinics, and the Government hospital in Aruba. In addition, the ICMPD (Introduction to Clinical Medicine and Physical diagnosis) course is designed to teach the history taking and physical examination skills, diagnostic reasoning and train the students for OSCE’s using standardized patients. In 2013, XUSOM launched its own “Standardized Patient Program” in Aruba. Since then, the Objective Structured Clinical Examination (OSCE) using standardized patients (SPs), is incorporated in each organ system course, to reinforce teaching and assessment of clinical skills training, and also to prepare students for the United States Medical Licensing Examination Step 2 Clinical Skills examination (USMLE Step 2CS). The use of the OSCE has grown in importance in medical education and assessment analysts are now labeling it as one of the most rational, effective, and dependable methods for assessing clinical performances.[1-5] It is one of the formats of competency-based assessment methods to evaluate not only medical knowledge, but also other core proficiencies, like practice-based learning and communication skills, which all are related to effective patient care.[2,6-8] The OSCE, with reference to Miller’s pyramid of assessment (Miller 1990) assess that a candidate is able to “show how” one would perform in simulated settings.[8] SP-based multiple station OSCEs are now a part of several high stakes examinations, including the Canadian Medical Council of Canada (MCC) qualifying examination and an examination for international medical graduates wishing to practice in Canada.[9,10] The National Board of Medical Examinations (NBME) also uses OSCE in the U.S. licensing examinations. Use of OSCEs in such high-stake examinations emphasizes the need of standard setting procedures to accurately assess the examinees during their training in medical schools. Standard setting methods are broadly categorized into norm-referenced or relative method and criterion-referenced method or absolute method. Norm-referenced or relative methods identify the cut-off score relative to performance of the group or top scoring examinees taking the examination.[11-14] Criterion-based or absolute methods identify cutoff scores based on the level of competence expected of students on the content being examined and are, thus, preferred for competence-based assessments like OSCEs.[15,16] This can further be categorized into test/examination-centered (eg, Angoff) methods and examinee-centered methods (eg, borderline group [BLG] and borderline regression).[17] With the requirement for standards to be defensible, evidenced and acceptable, absolute standards are generally preferred.[18] Medical schools commonly use examinee-centered standard setting methods such as Mean borderline group (BG) method and Borderline regression method (BLR).[13,19,20] An alternative method is test item centered, Modified Angoff standard setting method. In this method, the pass mark is based on item or station characteristics and varies according to the difficulty level of the station determined by the characteristics of the items on checklist rather than the examinees performance.[21] Until September 2018, XUSOM Aruba was applying an arbitrary cut off score of 70% as a passing score for OSCE. This decision was based on tradition, rather than on test content or examinee’s performance. It was difficult to provide a defensible explanation of how this 70% passing standard was set. Also, large variations in the performance of different cohorts, at XUSOM, were noticed while using the arbitrary 70% score as pass-fail criteria. The present study is proposed to address this problem by comparing 4 different standard setting methods with our traditional method, analyze the data and determine the method or combination of methods that would be most appropriate for the assessment of OSCE at XUSOM, Aruba. The purpose of this study was to compare the utility, feasibility and appropriateness of different standard-setting methods for setting pass-fail cut-off scores for internal OSCE examinations.

Methods

The current study is a descriptive study design conducted at XUSOM, Aruba. Basic Sciences students (year 1 and year 2) undertaking the final OSCEs in the Respiratory system (RS), Gastrointestinal system (GIS) and Cardiovascular system (CVS), during Spring and Summer 2019 semesters, participated in this study. In Spring 2019 semester, 33 students undertook the final Respiratory system OSCE, 33 students the GIS OSCE and 35 students the CVS OSCE. In Summer 2019, 22 students completed the Respiratory system OSCE, 20 students the GIS OSCE and 30 students the CVS OSCE (Table 1). Thus, a total of 173 students participated in this study. RS and GIS are taught to 3rd semester (Year 1) students separately at different times and CVS is taught to 4th semester (Year 2) students, therefore same set of students were administered with RS and GIS OSCE in Spring 2019 at different times, and these students were administered with CVS OSCE in Summer 2019. The students of 3rd semester (year 1) were administered with RS OSCE and GIS OSCE at different times during Summer 2019. The students for CVS OSCE in their 4th semester (Year 2) during Spring 2019 were unique (Table 1). Five full time faculty members with a minimum MD qualification were involved in the study. The project was approved by the IRB (Institutional Review Board) of XUSOM, Aruba. Informed consent and confidentiality of information about participants was obtained for the study.
Table 1.

Distribution of students across OSCE stations for 2 semesters with the number of students.

Spring 2019Summer 2019
Respiratory system (33 students)*Respiratory system (22 students)**
Gastrointestinal system (33 students)*Gastrointestinal system (20 students)**
Cardiovascular system (35 students)[#]Cardiovascular system (30 students)*

Same group (3 unique). **Same group (2 unique). #Unique.

Distribution of students across OSCE stations for 2 semesters with the number of students. Same group (3 unique). **Same group (2 unique). #Unique. A 15-station OSCE, using standardized patients, was administered to 173 students over 6 months. 5 stations each were held per system in a single circuit (total of 15 stations for 3 organ system) as shown below (Figure 1). Students rotated through the stations completing single circuit. Every student was examined on one-on-one basis by a single examiner.
Figure 1.

OSCE circuit.

OSCE circuit. An established blueprinting process was followed to ensure stations assessed a variety of appropriate domains of clinical skills such as knowledge, psychomotor and affective. Each station was of a 2-minute duration to assess communication skills and professionalism, 5 minutes for history taking, 6 minutes for physical examination skills and 2 minutes for interpretation of laboratory results and reaching most likely diagnosis. Each station was reviewed and validated by the faculty members prior to OSCE administration. The students were properly instructed, prior to the OSCE session, regarding the information related to the presenting problem, the task and time frame for completing the encounter and strict policy to be followed.[22] One full-time faculty member with several years’ experience of teaching ICMPD and administering OSCEs recruited other members of the study gradually over a period of time. To ensure consistency and fairness of scores, all the faculty involved were trained gradually via workshops for conducting OSCE’s during which they were clearly informed about the objectives, outcomes, roles and responsibilities and were allowed to shadow and observe. All the faculty members were involved in training the students for OSCE, designing the stations, creating rubrics for the OSCE stations based on the blueprint, in training of SP’s and grading the student’s performance during OSCE. The Station checklists were reviewed and validated by all members of faculty involved in the study. All SP’s were well trained for their consistent role to ensure that each student is presented with the same challenge.[23] The data analysis was done using ANOVA and t-test by using SPSS 20 version. The effect size was calculated using Cohen’s “d’ values for the pooled standard deviation. The effect size was deemed Large if d > 0.8, Medium if d > 0.5, and Small if d > 0.2.

Standard setting methods

Traditional method

Each examiner graded individual student performance using a station checklist comprising varying number of items per checklist based on the task assessed at each station with a minimum number of 5 items per checklist and a maximum number of 25 items per checklist. Based on the performance of the student, the score was converted to percentage. For each system, 5 OSCE checklists were used and checklist item criteria for communication skills and professionalism were common across all systems. The items for other clinical skills varied according to the clinical condition and organ system. Each student’s “System Total Score” was calculated by adding the 5 station checklist scores and dividing by 5 to produce an average system score for the student. Traditionally the arbitrary cut-off passing score of 70% had been applied to determine pass-fail.

Modified borderline group method (MBGM)

After completing the station checklists, the examiners also provided a global Likert rating (from 1 to 5) for every individual student, based on overall impression of the student’s performance, independent of the checklist score. Likert ratings included: 1 = Clear Fail, 2 = Borderline Fail, 3 = Borderline Pass, 4 = Clear Pass, and 5 = Excellent Performance. A linear regression analysis was conducted to ascertain whether station global Likert ratings correlated with the station checklist score. After establishing the relationship between Likert ratings and checklist scores, students obtaining Likert ratings between 2 (Borderline Unsatisfactory) and 3 (Borderline Satisfactory) at each station, were selected, to calculate a mean checklist score for the “Borderline Group” at each station.

Borderline regression method (BLR)

This method was conducted to predict checklist scores for a Borderline student. We used 2 borderline categories, 2 = Borderline Unsatisfactory and 3 = Borderline Satisfactory, with a mean Likert rating of 2.5 used as Borderline. For each of the 15 stations, we used a linear regression model in which the student’s checklist scores and global Likert rating scores were considered as dependent and independent variables, respectively. Then we calculated the checklist score cut-off on the regression equation for the global Likert rating cut-off set at 2.5.[21,24,25]

Modified Angoff method (MAM)

In this method, individual OSCE stations were scored by 6 selected Chairs of the clinical departments at XUSOM who also acted as judges. Each of the clinical chairs had more than 15 years of experience in teaching and assessing the students of year 3 and year 4 of the MD program and training the residents at their clinical departments. All the Clinical chairs had extensive expertise in conducting OSCEs, they were briefed about the purpose and steps of standard setting processes, followed by a brief discussion on qualities of a borderline (minimally competent) student. Each judge was asked to determine the probability of a borderline (minimally competent) student to perform the test item in each station correctly in the percentage from 0 to 100. Following the individual ratings, the judges displayed their ratings, then discussed the reasoning behind any discrepancies. Following the discussion, each judge again rated each station answering the same question. The judges’ estimates were averaged for each station and the mean of the averages was used as cut off scores.[26]

Relative method

Kaufman et al[13] used the best student’s performance as reference point, with such students generally well prepared for the examination and fluctuations in these students’ scores reflecting variations in exam difficulty. This method used the score that ranks at the 95th percentile and defined “passing” as a score that is equal to 60% of the 95th percentile. Similar methodologies of standard settings are shown to be more practical, and overcome certain disadvantages of criterion and norm-based methods.[27] But this definition resulted in zero failures in our study. We modified Kaufman’s method to define “passing” as the score that is equal to 70% of the 95th percentile score of the best students, which is consistent with our traditional cutoff standard of 70%.

Results

Table 2 summarizes the system total mean scores and percentage of students failing each system using the traditional method standard setting approach. The CVS and RS mean scores were significantly below the traditional standard of 70 (P < .05), and the CVS mean score was significantly lower than the GIS mean (P < .05).
Table 2.

Summary of traditional method standard setting approach.

System totalNumber of students (N)MeanStandard deviation95% Confidence intervalNumber of failures (below 70)% of failures (below 70)
CVS6564.407.1363 to 6754 out of 6583.1%
GIS5369.638.2667 to 7227 out of 5350.9%
RS5567.266.1066 to 6935 out of 5563.6%
(CI = Mean score ± 1.96 * Standard error of the mean for 95% CI) Standard error = S.D./√N Summary of traditional method standard setting approach. Pell et al[28] determined an acceptable R2 value to be above 0.50, and all values in our data, were above 0.50, indicating a strong linear relationship between the Likert rating and the checklist score for each station (Table 3). In our study, the R2 values are higher indicating that the checklist scores and the global scores show a strong positive correlation.
Table 3.

Borderline methods: R2 coefficient of determination.

Station R 2
10.97
20.58
30.78
40.92
50.94
60.97
70.88
80.93
90.91
100.97
110.98
120.77
130.87
140.86
150.95
Borderline methods: R2 coefficient of determination. Figures 2 to 4 depicting the average scores of the students for that particular system, they show consistent scatter plots as the individual station scores.
Figure 2.

Inter-grade discrimination for cardiovascular system.

Figure 4.

Inter-grade discrimination for respiratory system.

Inter-grade discrimination for cardiovascular system. Inter-grade discrimination for gastrointestinal system. Inter-grade discrimination for respiratory system. Table 4 shows the number of students identified as Borderline for each station, along with the Borderline Groups’ station checklist mean score. For each System OSCE, the sum of the 5 stations check list mean score was divided by 5 to determine the cut-off score. Failure rate was defined as the number/percent of students with a system total score lower than the cut-off score.
Table 4.

Modified borderline group method (MBGM): station means, standard deviation, and failure rate.

StationNumber of students with Likert 2 to 3Checklist mean scoreStandard deviationNumber of failures% of failures
Cardiovascular system (CVS)
13065.83
21963.65
33162.31
43266.28
54164.65
System total*64.541.6237 out of 6556.9%
Gastrointestinal system (GI)
61765.55
72167.55
81464.13
91970**
101764.42
System total*66.32.4518 out of 5334%
Respiratory system (RS)
112364.44
123265.39
133065.47
142569.16
153065.86
System total*66.061.8123 out of 5541.8%

Total mean score of respective organ system OSCE. **The Station 9 distribution of scores for the Borderline sample was severely negatively skewed. Therefore, the median (not the mean) is the most appropriate measure of central tendency and the median is reported instead of the Mean.

Modified borderline group method (MBGM): station means, standard deviation, and failure rate. Total mean score of respective organ system OSCE. **The Station 9 distribution of scores for the Borderline sample was severely negatively skewed. Therefore, the median (not the mean) is the most appropriate measure of central tendency and the median is reported instead of the Mean. Table 5 summarizes the predicted checklist score for a 2.5 global Likert rating at each station. For each system, the 5 predicted checklist scores were summed and divided by 5 to calculate a system total cut-off score. Failure rate was defined as the number/percent of students with a system total score lower than the cut-off score. Figures 2 to 4 represent the results of the Borderline Regression Method for each station by determining intergrade discrimination and indicating the relationship between checklist score and global Likert rating score on the slope of the regression line. Scores that fall along the slope of the regression line indicates checklist scores correlate well with the scores of Likert rating scale, which in turn reflects the validity of final score. Slopes of regression for Cardiovascular, Gastrointestinal and Respiratory systems indicate an average increase in checklist scores (y axis) corresponding to an increase of 1 grade on the global Likert rating scales (x axis).[25]
Table 5.

Borderline regression method (BLR): regression equation, borderline score, and failure rate.

StationRegression equation (Y = a + bX)Borderline (2.5) scoreNumber of failures% of failures
Cardiovascular system (CVS)
19.396X + 42.2365.72
29.777X + 40.4964.92
39.821X + 39.12263.67
410.115X + 0.30965.60
510.865X + 7.26964.43
System total*64.8737 out of 6556.9%
Gastrointestinal system (GIS)
69.651X + 41.29965.43
78.822X + 43.87165.93
89.308X + 41.54164.81
99.365X + 42.08965.50
109.914X + 40.11864.90
System total*65.3115 out of 5328.3%
Respiratory system (RS)
119.604X + 41.52565.54
128.886X + 43.32165.54
139.368X + 41.19964.62
148.212X + 46.93867.47
159.867X + 40.22564.89
System total*65.6120 out of 5536.4%

Mean checklist score of borderline students in respective organ system OSCE.

Borderline regression method (BLR): regression equation, borderline score, and failure rate. Mean checklist score of borderline students in respective organ system OSCE. Table 6 presents the MAM mean and standard deviation for each system, along with the failure rates, defined as number/percent of students with a system total score lower than the MAM cut-off score for the system.
Table 6.

Summary of judges’ estimates using MAM method.

SystemMeanStandard deviationNumber of failures% of failures
Cardio vascular55.883.856 out of 659.2%
Gastro intestinal59.744.586 out of 5311.3%
Respiratory57.912.674 out 557.3%
Summary of judges’ estimates using MAM method. Table 7 represents the 95th percentile for each system total score, along with the score that equals 70% of the 95th percentile and failure rates, defined as the number/percent of students who had a System Total score less than 70% of the 95th Percentile.
Table 7.

Relative method: percentile scores and failure rate.

SystemSystem total score that ranks at 95th percentile70% of 95th percentileNumber of failures% of failures
Cardio vascular75.92533 out of 654.6%
Gastro intestinal80.97574 out of 537.5%
Respiratory75.64531 out of 551.8%
Relative method: percentile scores and failure rate. Before analyzing the data to compare the different standard setting methods, the internal consistency of the OSCE across all the stations was assessed for all 173 students. The acceptable value of Cronbach’s alpha is 0.70 or above.[28] In our study, Cronbach’s alpha was 0.711, which is above the 0.70 acceptable value. This suggests that the checklist scores are internally consistent. As shown in Table 8, the “Alpha If Item Deleted” column estimates what Cronbach’s Alpha would be if we removed a specific station score. If a station’s “Alpha if deleted” value is lower than the total 0.711, it suggests reliability would decrease if we removed that station. In other words, any station with an “Alpha if Deleted” value lower than our total 0.711 is a station that strengthens our reliability because removing that station would weaken internal consistency.
Table 8.

Cronbach’s alpha if deleted data.

StationAlpha if deleted
10.705
20.698
30.687
40.734*
50.694
60.702
70.677
80.673
90.677
100.662
110.716*
120.709
130.697
140.705
150.695

“Alpha if deleted” value above the value of Cronbach’s alpha (0.711).

Cronbach’s alpha if deleted data. “Alpha if deleted” value above the value of Cronbach’s alpha (0.711). As shown in Table 8, 13 of the 15 stations performed well, with “Alpha if Deleted” values lower than 0.711. On the other hand, if we deleted Stations 4 and 11, Cronbach’s Alpha would increase above 0.711, which means that including Stations 4 and 11 weakens our internal consistency. However, the “Alpha if Deleted” values for stations 4 and 11 are not drastically above 0.711, so they do not severely detract from the reliability of checklist score. The effect size was calculated using Cohen’s “d” values for the pooled standard deviation of 5.57. A One-sample t-test shows that the CVS mean (M = 64.40, SD = 7.13) was significantly below the traditional standard of 70, t (64) = −6.33, P < .001. The effect size was “large” d = 0.84. Likewise, a One-sample t-test shows that the RS mean (M = 67.2, SD = 6.10) was significantly below the traditional standard of 70, t (54) = −3.32, P < .002. The effect size was “medium” d = 0.59. The GIS mean (M = 69.6, SD = 8.26) was not different than the traditional standard of 70, t (52) = −.31, P > .02. The effect size was “small” d = 0.05. For these multiple comparisons, alpha was adjusted from .05 to .02 to reduce the risk of a Type 1 error.

Comparison of system checklist means

A One-Way Between-Subjects ANOVA was conducted to compare the system mean scores, and the results were significant, F (2, 170) = 7.80, P < .001. Bonferroni post-hoc comparisons show that the CVS mean (M = 64.4, SD = 7.13) is significantly lower than the GIS mean (M = 69.6, SD = 8.26). We are 95% Confident that the CVS mean is between 2 to 8.5 points lower than the GIS mean. The RS mean was are not different than either CVS or GIS. Table 9 summarizes the cut-off score and % of failed students for each system, using each of the five standard setting methods.
Table 9.

Summary: comparison of standard setting methods.

Cut-off score% of failed students
Traditional method
CVS64.483.1
GIS69.350.9
RS67.663.6
Relative method
CVS53.04.6
GIS57.07.5
RS53.01.8
Modified Angoff method
CVS55.99.2
GIS59.711.3
RS57.97.3
Borderline group method
CVS64.556.9
GIS66.334.0
RS66.141.8
Borderline regression method
CVS64.956.9
GIS65.328.3
RS65.636.4
Summary: comparison of standard setting methods.

Discussion

The results indicate that the traditional arbitrary score of 70 had the highest failure rate, with the majority of students failing all 3 organ systems. The MAM and Relative methods yielded the lowest failure rates, which were typically less than 10% for each system. Failure rates for the Borderline methods ranged from 28 to 57% across the systems. Mean scores for the CVS was the lowest, and mean scores of GIS was relatively high. One of the reasons that CVS scores were low could be attributed to the CVS station checklists being more challenging with more items assessing clinical reasoning skills. On the contrary, high GIS scores could be because the items in the checklist of the stations in GIS were uniformly process based requiring relatively less clinical reasoning. Therefore, to correct this disparity we proposed that the checklists of these 2 systems be reviewed and be comparable and standardized with regards to communications skills items and clinical reasoning items in future. We also proposed that items of CVS checklist be made more liberal encompassing more items to test the basic aspects of history taking in CVS rather than items assessing advanced clinical judgement, and to incorporate more items that assess the clinical reasoning of examinee in the checklists of GIS. Kaufman et al[13] reported that Angoff and borderline methods were shown to provide a reasonable and defensible approach to standard setting and were of practical value when used by non-psychometricians in medical schools.[26] In contrast, Kramer et al examined standard setting in postgraduate general practice training, identifying that the Borderline Regression Method (BRM) was more credible and acceptable than the modified Angoff method (MAM). Kramer et al[29] used 84 examiners and a significant number of them also performed the modified Angoff method. These conflicting findings may be the results of some known difficulties with the Angoff method. Verheggen et al[30] demonstrated considerable variation between judges, especially when judges had less expertise in certain item areas. In our study, though we used only 6 clinical chairs as judges compared to larger numbers used in other similar studies, the results of the modified Angoff method (MAM) appeared to be credible and acceptable. The judge’s clinical experience, subject mastery and involvement in supervising OSCEs might have led them to correctly answer all the items in the checklists used to determine the scores of a borderline student. This may be the reason, which explains achievement of an acceptable cut off score. In a study by Dwyer et al[21] similar observation was made about modified Angoff method (MAM) to set acceptable and credible cut-scores compared to the borderline method and Borderline Regression Method (BRM). Boursicot et al[20] observed that the Borderline method is more consistent in determining the pass score than the Angoff method. Wood et al[24] reported that a cut score derived from a Borderline Regression Method (BRM) was more accurate than 1 derived using the modified borderline group method, supporting the finding of our study as well. Hejri et al[25] also reported that Borderline Regression Method (BRM) is much more convenient and less resource consuming compared to other procedures like Angoff. Also, BRM has the advantage of generating a number of indices that are useful in measuring the validity of the OSCE. Considering the fact that BRM is widely used as a standard setting method, assessing its reliability is of paramount importance. In our study, both Modified Angoff method (MAM) and Borderline Regression Method (BRM) were shown to be reliable by which there were consistently similar cut-off scores across different organ systems, thereby providing way to decide acceptable cut off scores. The pass/fail standard can be reliably set before the OSCE by using Modified Angoff method (MAM), which would be further useful in the setting of competency based medical education (CBME).[31-33] Borderline Regression Method (BRM) provides objective analysis, and statistical approach to decide about accurate cut off score and to ensure validity of the OSCE by assessing degree of correlation (R2) between the checklist score and the overall global rating score. Kaufman et al[13] showed that the relative and traditional methods gave inconsistent results which was similar to the findings of our study. The findings of our study discourage the use of absolute / traditional method and relative method to determine pass-fail cut off score. Further, according to our study modified borderline group method (MBGM) seems not to be reliable when applied to a small scale OSCEs. It appears that none of the methods of standard setting are perfect when used alone. The standard setting results in our study have either a very high failure rate or very low cut-off score. This creates a disparity in the assessment of the borderline group of students, whose result depends on the cut-off scores. If the cut-off scores are too lenient then it gives advantage to the weak students who have just followed the process scale (communication skills), but did not have any clinical intuition in approaching a case. If the mean cut-off scores are too strict, then it becomes a disadvantage to the set of students who completed a difficult station. Therefore, to minimize this disparity we propose, to use a combination of standard setting methods by combining 2 methods that could establish a reliable cut-off score and determine an acceptable percent of students failing an examination. It was determined that a combination of BRM and MAM could be practical. In MAM, subject experts determine the cut off score relying on their subjective and professional judgment based on the characteristics of test items/station. In BRM, the cutoff scores are determined by the assessment of examinee’s actual performance by expert examiners thereby providing an objective and statistical approach to determine an acceptable cut off score of borderline students. Also, in our study analysis of R2 coefficient and intergrade discrimination values obtained by BRM ensures the quality of the overall OSCE and consistency of the examiner grading across all stations during OSCE.[27] The MAM average across all 3 systems is 58 and the BRM average is 62, so the combined average is 60. This gives us acceptable and reliable minimum score that a student has to achieve, to be able to pass the OSCE and also helps us differentiate a borderline pass vs. a borderline fail. Yousef et al[34] similarly observed that combining and averaging 2 standard settings, Angoff and Hofstee methods yielded a desirable higher cutoff passing scores than the fixed arbitrary passing score of 60% that was used in their school. Confounding factors creating disparity in the scores such as knowledge, attitude and practices of the Year 1 and Year 2 students toward applying concepts of clinical reasoning in OSCE’s, which may require revisions in other phases / steps of the examination cycle.

Conclusion

None of the standard setting methods, when used alone was of pragmatic value to determine an acceptable and reliable cut off score. However, using a combination of Modified Angoff method (MAM) with Borderline Regression Method (BRM) seems to produce a reliable and valid determination of cut off score. Further studies, in high-stake clinical examinations, utilizing larger number of judges and OSCE stations are recommended to reinforce the validity of combination of multiple methods for standard settings.
  27 in total

1.  Standards and reliability in evaluation: when rules of thumb don't apply.

Authors:  J J Norcini
Journal:  Acad Med       Date:  1999-10       Impact factor: 6.893

Review 2.  An objective structured clinical examination for the licentiate: report of the pilot project of the Medical Council of Canada.

Authors:  R Reznick; S Smee; A Rothman; A Chalmers; D Swanson; L Dufresne; G Lacombe; J Baumber; P Poldre; L Levasseur
Journal:  Acad Med       Date:  1992-08       Impact factor: 6.893

3.  How to measure the quality of the OSCE: A review of metrics - AMEE guide no. 49.

Authors:  Godfrey Pell; Richard Fuller; Matthew Homer; Trudie Roberts
Journal:  Med Teach       Date:  2010       Impact factor: 3.650

4.  Standard setting in a small scale OSCE: a comparison of the Modified Borderline-Group Method and the Borderline Regression Method.

Authors:  Timothy J Wood; Susan M Humphrey-Murto; Geoffrey R Norman
Journal:  Adv Health Sci Educ Theory Pract       Date:  2006-05       Impact factor: 3.853

5.  Comparison of students' performance in the traditional oral clinical examination and the objective structured clinical examination.

Authors:  Talal M Bakhsh; Abdulrahman M Sibiany; Faisal M Al-Mashat; Abdulrahman A Meccawy; Fatma K Al-Thubaity
Journal:  Saudi Med J       Date:  2009-04       Impact factor: 1.484

6.  Standard Setting Methods for Pass/Fail Decisions on High-Stakes Objective Structured Clinical Examinations: A Validity Study.

Authors:  Naveed Yousuf; Claudio Violato; Rukhsana W Zuberi
Journal:  Teach Learn Med       Date:  2015       Impact factor: 2.414

7.  Challenging the arbitrary cutoff score of 60%: Standard setting evidence from preclinical Operative Dentistry course.

Authors:  M K Yousef; L Alshawwa; A Tekian; Y S Park
Journal:  Med Teach       Date:  2017-01-25       Impact factor: 3.650

8.  Objective structured clinical/practical examination (OSCE/OSPE).

Authors:  N Ananthakrishnan
Journal:  J Postgrad Med       Date:  1993 Apr-Jun       Impact factor: 1.476

9.  Students' performance in the different clinical skills assessed in OSCE: what does it reveal?

Authors:  Joong Hiong Sim; Yang Faridah Abdul Aziz; Azura Mansor; Anushya Vijayananthan; Chan Choong Foong; Jamuna Vadivelu
Journal:  Med Educ Online       Date:  2015-02-18

Review 10.  Assessing Communication Skills of Medical Students in Objective Structured Clinical Examinations (OSCE)--A Systematic Review of Rating Scales.

Authors:  Musa Cömert; Jördis Maria Zill; Eva Christalle; Jörg Dirmaier; Martin Härter; Isabelle Scholl
Journal:  PLoS One       Date:  2016-03-31       Impact factor: 3.240

View more
  3 in total

1.  Comparison of OSCE performance between 6- and 7-year medical school curricula in Taiwan.

Authors:  Jr-Wei Wu; Hao-Min Cheng; Shiau-Shian Huang; Jen-Feng Liang; Chia-Chang Huang; Ling-Yu Yang; Boaz Shulruf; Ying-Ying Yang; Chen-Huan Chen; Ming-Chih Hou; Wayne Huey-Herng Sheu
Journal:  BMC Med Educ       Date:  2022-01-04       Impact factor: 2.463

2.  Measuring the Effect of Using a Borderline Students' Characteristics Model on Reliability of Objective Structured Clinical Examination.

Authors:  Rabab A Abed; Shimaa E Elaraby
Journal:  Cureus       Date:  2022-05-20

3.  A structured multimodal teaching approach enhancing musculoskeletal physical examination skills among undergraduate medical students.

Authors:  Abdulaziz Z Alomar
Journal:  Med Educ Online       Date:  2022-12
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.