Literature DB >> 33102347

Psychometric analysis of multiple-choice questions in an innovative curriculum in Kingdom of Saudi Arabia.

Karim Eldin M A Salih1,2, Abubakar Jibo3, Masoud Ishaq2, Sameer Khan4, Osama A Mohammed5, Abdullah M Al-Shahrani3, Mohammed Abbas1,2.   

Abstract

BACKGROUND AND AIMS: Worldwide, medical education and assessment of medical students are evolving. Psychometric analysis of the adopted assessment methods is thus, necessary for an efficient, reliable, valid and evidence based approach to the assessment of the students. The objective of this study was to determine the pattern of psychometric analysis of our courses conducted in the academic year 2018-2019, in an innovative curriculum.
METHODS: It was a cross-sectional-design study involving review of examination items over one academic session -2018/2019. All exam item analysis of courses completed within the three phases of the year were analyzed using SPSS V20 statistical software.
RESULTS: There were 24 courses conducted during the academic year 2018-2019, across the three academic phases. The total examination items were 1073 with 3219 distractors in one of four best option multiple choice questions (MCQs). The item analysis showed that the mean difficulty index (DIF I) was 79.1 ± 3.3. Items with good discrimination have a mean of 65 ± 11.2 and a distractor efficiency of 80.9%. Reliability Index (Kr20) across all exams in the three phases was 0.75. There was a significant difference within the examination items block (F = 12.31, F critical = 3.33, P < 0.05) across all the phases of the courses taken by the students. Similarly, significant differences existed among the three phases of the courses taken (F ratio = 12.44, F critical 4.10, P < 0.05).
CONCLUSION: The psychometric analysis showed that the quality of examination questions was valid and reliable. Though differences were observed in items quality between different phases of study as well as within courses of study, it has generally remained consistent throughout the session. More efforts need to be channeled towards improving the quality in the future is recommended. Copyright:
© 2020 Journal of Family Medicine and Primary Care.

Entities:  

Keywords:  Item analysis; Saudi Arabia; UBCOM; courses; phases of study

Year:  2020        PMID: 33102347      PMCID: PMC7567208          DOI: 10.4103/jfmpc.jfmpc_358_20

Source DB:  PubMed          Journal:  J Family Med Prim Care        ISSN: 2249-4863


Introduction

Curriculums are guides used by teachers in schools to assist in the education of students. It contains objectives, activities units and suggested materials to enhance learning.[1] Curricular innovation is a managed process of development whose main products are teaching (and testing) materials, methodological skills, and pedagogical values perceived as new by potential adopters.[2] It is a willed intervention, which results in the development of ideas, practices, or beliefs, that are fundamentally new. In innovated, integrated curriculum, designing multiple-choice questions (MCQ) for assessments is a complicated and time-consuming process.[3] MCQs are the most commonly used tool for assessment of students in different courses offered at undergraduate and postgraduate levels, and capable of yielding examination items from the contents of the taught courses.[4] These items, when critically analyzed, provides feedback to both tutors and students on performance on each test item. Psychometric analysis of any test is defined sequences of events to collect data from a test to determine its quality.[3] One of the importance of item analysis is to know the reliability or consistency of the test administered.[35] This will ensure accountability to the community by providing competent graduates. The reliability of a test indicates its consistency, homogeneity and ultimately acceptability as a tool of measurement.[5] In item analysis, item difficulty and its ability to discriminate between students who knows and those who do not know determine the quality of the examination.[67] Providing a reliable test with reasonable difficulty will result in a type of assessment that can derive learning.[8] According to Ebel,[8] in a classical test theory item analysis, discrimination index (DI) of greater than 0.2 is acceptable. However, other workers suggested that any value above 0.15 is acceptable.[89] The difficulty index (DIF I) is determined by the number of the candidates who got the answer right over the total number of the students. A reasonable test should have difficulty index (DIF I) in the range of 50-80%.[10] Some authors consider DIF I above 80% as high implying that the questions are easy. On the other hand, low DIF I (less than 30%) means that the questions are difficult with a pressing need to improve the quality of the test item.[111213] Discrimination index (DI), also called biserial point correlation (PBS), describes the ability of an item to distinguish between high and low scorers.[14] It ranges between -1.00 and + 1.00. It is expected that the high-performing students select the correct answer for each item more often than the low-performing students. If, however, the low performing students got a specific item correct more often than the high scorers, then that item has a negative DI (between -1.00 and 0.00).[15] The difficulty and discrimination indices are often reciprocally related. However, this may not always be true. Questions having high DI-value (more straightforward questions) tend to discriminate poorly; conversely, questions with a low DI-value (harder questions) are considered to be good discriminators.[16] Discrimination index of 0.40 and above is excellent, 0.30-0.39 is reasonably good, 0.20-0.29 is marginal items (i.e. subject to improvement), and 0.19 or less is poor items (i.e. to be rejected or improved by revision).[1217] A general indicator of test quality is the reliability estimate usually reported on the test analysis printout. Referred to as KR-20 or Coefficient Alpha, it reflects the extent to which the test would yield the same ranking of examinees if re-administered with no effect from the first administration.[18] Reliability (R) range from 0.7-1.0 is considered by many authors as excellent and acceptable.[131920] A distracter efficiency (DE) is the list of distracters that distract and in an MCQ of three distractors it is 100%, 66%, 33% or 0%, if all the three distractors are distracting, tow distractors chosen, one distractor chosen or all distractors not chosen. A functional distractor is distractors that has been attempted by at least 5% or more of the students.[921] According to Ebel and Downing, only 38% of distractors on the tests are eliminated because < 5% of students select them.[8] He reported that the percentage of items with three functioning distractors in most tests ranged from only 1.1 to 8.4% of all items.[22] The ultimate Goal and objective of each medical institutions should primarily directed to provision of evidence based care of the patient, however proper assessment by conducting high quality psychometric analysis of our assessment will ensure this. In fact gratifying advantage of multiple choice questions (MCQs), is its ability to provide immediate feedback to all partners. This feedback will be maximally utilized if it combined with psychometric analysis. As it has been observed by many authorities psychometric analysis is an effective tools in in deciding what number of options, keeping questions in bank and comparing students achievement.[23] Amani et. al. (2020) used psychometric analysis to compare quality of MCQs designed by residents in radiology program and that design by teacher.[23] To ensure fair response from the examinees, many factors should be considered: on the top of these factors is the reliability of the questions and the test, the quality of the questions, the validity of the test and finally the language back ground of the examinee. Psychometric analysis will provide some answers to these questions by tell us how many option is needed for those who uses English as second language and can alert us to possible areas for improvement.[23] The rational of this study is to through some light across MCQs adopted for assessment as an effective tool, to draw the attention of teachers and administrations to the importance of unflawed MCQs and checking it's psychometric pattern.

The Objective

The objectives of our study is to determine the psychometric analysis of our courses conducted in the academic year 2018-2019, in an innovative curriculum.

Methods

Study design

The study is cross-sectional by design.

Site of the study

The study site was the College of Medicine, University of Bisha. The college was recently established to graduate competent doctors for the Kingdom.

Description of academic and assessment process

There are three phases integrated longitudinally and horizontally, with a total duration of six years study period. The curriculum adopts as methods of assessment, MCQs assessment, best of four, structured short answer questions, objective structured practical examination (OSPE), and objective structured clinical examination (OSCE). Quality chain for process of assessment is achieved through students’ assessment committee (SAC), departmental meetings, and College Board. It involves review of exam items held within the 2018/2019 academic session at the college.

Data collection

Data was collected from the examination office for six weeks by five-trained research assistants.

Data analysis

The data on exam item analysis was generated by an optic reader machine used in marking the MCQ, the Apperson Datalink 3000 manufactured by Apperson.com, USA. The research assistants that participated are academic staff of the college-trained on data extraction and review of exam items analysis. Tool for data collection was a semi-structured questionnaire that was earlier validated by a pretest. Adjustments were made to the tool to capture all information required to address the specific objectives of the study. The Questionnaire captured essential items of the examinations held namely difficulty index, discrimination index, point biserial statistics, discriminator indices and efficiency as well as the test reliability (Kruder-Richardson-20). The reliability through Kuder-Richardson formula 20 (KR-20), Difficulty and the Discrimination Indexes, and the distractors functionality were considered for each question.[2324] MCQ data was entered into Excel sheets and transferred to SPSS version 20 for analysis. Categorical variables were presented in the form of frequency and percentages. Tests of association were used to find relationships between variables of interest. The three phases were compared to detect any difference existing between and within the phases using a two- way analysis of variance (ANOVA). Similarly, differences in the questions within and between courses was carried out using ANOVA. F distribution statistics was used to determine variations between and within these two factors of interest, with F ratios and critical values determined. Significant differences are observed where the P values are less than 0.05. The ethical clearance was sought from the College Ethical Committee of the University of Bisha, College of Medicine to address the issues of concern.

Results

The total number of courses taken in the three phases of the session was 24, with first phase having 33.3% of the courses, second phase constituting 45.8% of the courses and the third phase having 20.8%. The total number of exam items was 1073 in these three phases. Of these, Phase one constituted 27.5% (n = 323), Phase two items were 45.2% (n = 530) and Phase three 18.8% (n = 220). Each question was made up of multiple-choice answer questions type A with three distractors and a key answer to the questions. The total number of distractors in 1073 MCQ questions was 3219 [Table 1]. Analysis of the exam items taken for the courses across the three phases of the session revealed that 11.4% of the exams questions during the session under review were tough (<30%). The mean difficulty across the three phases was 11.6 ± 1.8SD. Exams items that were very easy (ease index >85%) were16.8% of the questions with mean ease index score of 16.6 ± 5.1. The proportion of exams questions that were within acceptable DIF I (30-85%) was 71.9% with a mean of 71.9 ± 3.3 [Figure 1]. The discriminatory index (DI) showed that 24.2% of the questions poorly discriminates (DI <0.15) and have a mean of 25.1 ± 10.7SD. Similarly, the mean of the items with good discrimination was 65.2 ± 11.2 across all the three phases. The Point Biserial Statistics showed that 38.7% of the exam questions were of poor construction (Pbsr< 0.2) [Table 1]. Distractor indices analyzed showed 6.5% of the questions had three nonfunctioning distractors (3nFD), 16.5% had two nonfunctioning distractors (2nFD), and 33.7% had one nonfunctioning distractor (1nFD), in the session under review. The proportion of exam items with nonfunctioning distractors across all the three phases was 19.1%. The distractor efficiency (DE) observed within the 1073 items was 80.9% [Table 1]. Reliability index Kr20 across the three phases ranges from 0.5-0.8, and the mean KR 20 index for the exam items in the three phases is 0.754 [Table 1]. One-tenth (10%) of the questions in Phase 3 negatively discriminate students’ scores. About one-third of the questions (28.6%) have zero discrimination as shown in Figure 2. The results of the ANOVA across the three phases are as shown in Table 2. F ratio for the rows (exam questions) with df = 2 was 12.32 which is higher than the F critical value of 3.23. A significant difference (P < 0.05) was seen within the problem questions across all phases of the courses taken by the students. Similarly, the F ratio for the columns (phases) was 12.44, which is higher than the F critical value (4.10). Thus, rejecting the null hypothesis and interpreted as having significant difference existing between the three phases of the courses (P < 0.05) [Table 3].
Table 1

Summary of all the phases combined

CoursePhase onePhase twoPhase threeTOTAL



123456789101112131415161718192021222324
Ques3040504050304330706050506060405030606030303030401073
Dif Q25412936396581253411901356122 (11.4%)
Eas Q53835632159790912152231278692180 (16.8%)
Negd32151000333070151190115558 (5.4%)
1nFD1614121017151011291619191322131910152611157716362 (33.7%)
2nFD724762371711149458115217655105181 (16.9%)
3nFD3232251135130651001312110070 (6.5%)
Rel.0.60.70.80.80.80.70.90.80.90.90.80.90.90.80.80.70.80.80.70.80.80.60.50.7

Key: Dif Q- Difficult questions, Eas Q- Easy questions, Negd -Negatively discriminating questions, 1nFD- 1 nonfunctioning distractor, 2nFD-- 2 nonfunctioning distractor,3nFD - 3 nonfunctioning distractor, Rel- Reliability Kr index, Dist- MCQ Distractors

Table 1: Shows that more than a third of the questions (33.7%) have one non-functional distractor, 16.9% have two non-functional distractors, and 6.5% of the total questions have three non-functional distractors. The reliability index Kr20 across the three phases ranges from 0.5-0.8.

Figure 1

The figure shows examination difficulty index observed across the three phases of the courses taken. The mean DIF I was 71.8% across the three phases

Figure 2

One-tenth (10%) of the questions in phase three negatively discriminate students’ scores. about one-third of the questions (28.6%) have zero discrimination as observed

Table 2

All phases combined

Problem QuestionsPhase 1Phase 2Phase 3Row TotalRow average
No of diff Q44542412240.67
No of easy Q351014418060
No of neg discr Q1224215719
1 NFDs10517582362120.67
2 NFDs381053818160.33
3 NFDs194657023.33
Total253505214972
Column average42.1784.1735.67

Key: No of Diff Q- Difficult questions, No of Easy Q- Easy questions, No of Neg -Negatively discriminating questions, 1nFD- 1 nonfunctioning distractor, 2nFD--2 nonfunctioning distractor,3nFD - 3 nonfunctioning distractor, Rel- Reliability Kr index, Dist- MCQ Distractors

A two-way analysis of variance was done to determine the differences between the three phases of study as well as differences in the problem questions answered during the three phases.

Table 3

Anova: Two-factor without replication

SUMMARYCountSumAverageVariance
Diff Q312240.66667233.3333
easy3180601281
Neg DSC3571939
1nFD3362120.66672346.333
2nFD318160.333331496.333
3nFD37023.33333434.3333
P1625342.166671093.367
P2650584.166672990.967
P3621435.66667702.6667
ANOVA

Source of VariationSSdfMSFPF crit

Rows20591.3333354118.26712.316620.0005183.325835
Columns831724158.512.436950.0019394.102821
Error3343.66666710334.3667
Total3225217

The ANOVA (two way) results shows that the F ratio for the rows (exam questions) with df=2was observed to be 12.32 is higher than the F crit value of 3.23. It means a significant difference (P < 0.05) exist within the problem questions block across all the phases of the courses taken by the students. Similarly, the F ratio for the columns (phases) is 12.44, which is higher than the F critical value (4.10). Thus, rejecting the null hypothesis and interpreted as significant difference existing between the exam items in the three phases of the courses (P < 0.05).

No Q- number of questions. SS= sum square. Diff Q-Difficult questions. df =degree of freedom. Easy Q- easy questions. MS=Mean square. Neg DSCR-Negative discrimination question. F= F ratio. 81NFD- 1 non-functional distractor. F crit=F critical value. 2NFD- 2 non-functional distractor. 3NFD-- 1 non-functional distractor. P1-Phase one. P2-Phase two. P3- Phase three

Summary of all the phases combined Key: Dif Q- Difficult questions, Eas Q- Easy questions, Negd -Negatively discriminating questions, 1nFD- 1 nonfunctioning distractor, 2nFD-- 2 nonfunctioning distractor,3nFD - 3 nonfunctioning distractor, Rel- Reliability Kr index, Dist- MCQ Distractors Table 1: Shows that more than a third of the questions (33.7%) have one non-functional distractor, 16.9% have two non-functional distractors, and 6.5% of the total questions have three non-functional distractors. The reliability index Kr20 across the three phases ranges from 0.5-0.8. The figure shows examination difficulty index observed across the three phases of the courses taken. The mean DIF I was 71.8% across the three phases One-tenth (10%) of the questions in phase three negatively discriminate students’ scores. about one-third of the questions (28.6%) have zero discrimination as observed All phases combined Key: No of Diff Q- Difficult questions, No of Easy Q- Easy questions, No of Neg -Negatively discriminating questions, 1nFD- 1 nonfunctioning distractor, 2nFD--2 nonfunctioning distractor,3nFD - 3 nonfunctioning distractor, Rel- Reliability Kr index, Dist- MCQ Distractors A two-way analysis of variance was done to determine the differences between the three phases of study as well as differences in the problem questions answered during the three phases. Anova: Two-factor without replication The ANOVA (two way) results shows that the F ratio for the rows (exam questions) with df=2was observed to be 12.32 is higher than the F crit value of 3.23. It means a significant difference (P < 0.05) exist within the problem questions block across all the phases of the courses taken by the students. Similarly, the F ratio for the columns (phases) is 12.44, which is higher than the F critical value (4.10). Thus, rejecting the null hypothesis and interpreted as significant difference existing between the exam items in the three phases of the courses (P < 0.05). No Q- number of questions. SS= sum square. Diff Q-Difficult questions. df =degree of freedom. Easy Q- easy questions. MS=Mean square. Neg DSCR-Negative discrimination question. F= F ratio. 81NFD- 1 non-functional distractor. F crit=F critical value. 2NFD- 2 non-functional distractor. 3NFD-- 1 non-functional distractor. P1-Phase one. P2-Phase two. P3- Phase three

Discussion

The students’ performance in the MCQ of these 24 courses was used to determine the difficulty index, discrimination index and non-functional distractors or study evaluates how the MCQ differentiate between student's performance, within the test items in each of the courses and across the three phases of study. In this study, the DIF I of the items was 71.9% similar to what was reported by other studies,[3102526] where 61-80% of the items are within acceptable range. However, our study revealed the difficulty and ease indices in the three phases are lower than those reported by these studies. The discriminatory index (DI) shows up to a quarter of our assessment's items (25.1%) poorly discriminates between good and poor students (DI <0.15). Some authors[910] reported range of 14-17% of exam questions were poorly discriminating compared to about a quarter (25.1%) seen in our study. This difference could be partly due to the variation in the cut-off score adopted for the studies and partly ascribed to the different tools used for the study. About sixty-five per cent of the items showed excellent discrimination (>0.2). The DI index is similar to that reported by Rao et al.[27], where 60% of the items were good discriminators. Other authors[2829] have reported that DI of 0.2 is acceptable and would discriminate between weak and good students. We observed a distractor efficiency of 80.9% and non-functioning distractors accounting for 19.1% of questions across all phases. Two-third of the questions have one non-functional distractors, 16.9% have two and less than a tenth (6.5%) have three non-functional distractors. Tarrant in Hong Kong[21] reported similar to that where 13.8% of the total items he tested had only three functioning distractors, where 70% of the items had one or two functional distractors.[15] Other authors documented more than 66% of items showed non-functional dissectors [Table 3]. There was significant difference in difficulty factor, discrimination indices, and reliability and distractors functionality between different phases. The probable reasons for these results could be meticulous internal regulations adopted by the SAC, departments and course coordinators before and after conduction of the examinations. Weekly feedback to tutors is known to be essential in training,[30] this is made possible through panel discussions and other activities in our innovative curriculum. It improves the quality of training and understanding on the part of the students. It is obvious from this study and other studies adherence to restrict MCQs guidelines in general, beside application of cover hand test i.e. to anticipate the true answer even without looking at the options can provide better psychometric analysis regardless of option numbers.[24] To the knowledge of the authors unfocussed questions in general and which of the fallowing types of questions in particular will affect the quality of psychometric analysis.

Conclusion

Psychometric analysis of exam items showed that the quality of examination questions was valid and reliable. Variations in items quality have been observed between different phases of study as well as within courses of study that the quality of the exam items has generally remained consistent throughout the session.

Recommendation

Psychometric analysis is urgently needed so as to determine area of improvement and to build reliable MCQs bank.

Limitation

The data was collected from one institute during one academic year and numbers of students were less.

Strength

The work addresses the determination of the MCQs questions in an innovative curriculum.

Financial support and sponsorship

Nil.

Conflicts of interest

There are no conflicts of interest.
  10 in total

1.  Evaluating assessment: the missing link?

Authors:  S L Fowell; L J Southgate; J G Bligh
Journal:  Med Educ       Date:  1999-04       Impact factor: 6.251

2.  Analysis of one-best MCQs: the difficulty index, discrimination index and distractor efficiency.

Authors:  Mozaffer Rahim Hingorjo; Farhan Jaleel
Journal:  J Pak Med Assoc       Date:  2012-02       Impact factor: 0.781

3.  Evaluating and improving multiple choice papers: true-false questions in public health medicine.

Authors:  R A Dixon
Journal:  Med Educ       Date:  1994-09       Impact factor: 6.251

4.  Evaluation of vignette-type examination items for testing medical physiology.

Authors:  R G Carroll
Journal:  Am J Physiol       Date:  1993-06

5.  An assessment of functioning and non-functioning distractors in multiple-choice questions: a descriptive analysis.

Authors:  Marie Tarrant; James Ware; Ahmed M Mohammed
Journal:  BMC Med Educ       Date:  2009-07-07       Impact factor: 2.463

6.  Comparison in the quality of distractors in three and four options type of multiple choice questions.

Authors:  Nourelhouda A A Rahma; Mahdi M A Shamad; Muawia E A Idris; Omer Abdelgadir Elfaki; Walyedldin E M Elfakey; Karimeldin M A Salih
Journal:  Adv Med Educ Pract       Date:  2017-04-10

7.  Item and Test Analysis to Identify Quality Multiple Choice Questions (MCQs) from an Assessment of Medical Students of Ahmedabad, Gujarat.

Authors:  Sanju Gajjar; Rashmi Sharma; Pradeep Kumar; Manish Rana
Journal:  Indian J Community Med       Date:  2014-01

8.  Analysis of use of a single best answer format in an undergraduate medical examination.

Authors:  Fahmi Ishaq El-Uri; Naser Malas
Journal:  Qatar Med J       Date:  2013-11-01

9.  Adding to the debate on the numbers of options for MCQs: the case for not being limited to MCQs with three, four or five options.

Authors:  Mike Tweed
Journal:  BMC Med Educ       Date:  2019-09-14       Impact factor: 2.463

10.  Inclusion of MCQs written by radiology residents in their annual evaluation: innovative method to enhance resident's empowerment?

Authors:  Nadia Amini; Nicolas Michoux; Leticia Warnier; Emilie Malcourant; Emmanuel Coche; Bruno Vande Berg
Journal:  Insights Imaging       Date:  2020-01-23
  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.