Literature DB >> 24639820

Evaluation of Multiple Choice and Short Essay Question items in Basic Medical Sciences.

Mukhtiar Baig1, Syeda Kauser Ali2, Sobia Ali3, Nighat Huda4.   

Abstract

OBJECTIVES: To evaluate Multiple Choice and Short Essay Question items in Basic Medical Sciences by determining item writing flaws (IWFs) of MCQs along with cognitive level of each item in both methods.
METHODS: This analytical study evaluated the quality of the assessment tools used for the first batch in a newly established medical college in Karachi, Pakistan. First and sixth module assessment tools in Biochemistry during 2009-2010 were analyzed. Cognitive level of MCQs and SEQs, were noted and MCQ item writing flaws were also evaluated.
RESULTS: A total of 36 SEQs and 150 MCQs of four items were analyzed. The cognitive level of 83.33% of SEQs was at recall level while remaining 16.67% were assessing interpretation of data. Seventy six percent of the MCQs were at recall level while remaining 24% were at the interpretation. Regarding IWFs, 69 IWFs were found in 150 MCQs. The commonest among them were implausible distracters (30.43%), unfocused stem (27.54%) and unnecessary information in the stem (24.64%).
CONCLUSION: There is a need to review the quality including the content of assessment tools. A structured faculty development program is recommended for developing improved assessment tools that align with learning outcomes and measure competency of medical students.

Entities:  

Keywords:  Assessment; Item analysis; MCQ; SEQ

Year:  2014        PMID: 24639820      PMCID: PMC3955531          DOI: 10.12669/pjms.301.4458

Source DB:  PubMed          Journal:  Pak J Med Sci        ISSN: 1681-715X            Impact factor:   1.088


INTRODUCTION

Assessment is an essential part of the learning process in education. Students perceive it as a dominant motivator to direct and drive their learning.[1] The method of assessment determines the approach of students towards learning. Students’ are inclined to espouse a surface approach when assessment emphasis is on recall of factual knowledge and students are more likely to adopt a deep approach[2] if assessment demands higher levels of cognitive abilities. The approach to learning is a dynamic characteristic and is always modified according to students’ perceptions of the learning environment.[3] It has been reported that one of the most important factor influencing students’ choice of learning approach is the way how assessment is being conducted.[4]-[6] Multiple methods of assessment namely MCQs, SEQs, OSPE and VIVA are commonly used to assess Basic Science knowledge in undergraduate medical education in Pakistan. Multiple choice questions (MCQs) are the most frequently used type of tests deployed on their own or in combination with other types of test tools for assessment. Moreover, MCQs are appropriate for measuring knowledge, comprehension and could be designed to measure application and analysis.[7] MCQs are being used increasingly due to their higher reliability, validity, and ease of scoring.[8],[9] Essay-type assessment is a sensitive test requiring students not only to recall facts but also to use higher-order cognitive skills.[10] Essay questions though time consuming provides a unique evaluation tool particularly suited for the undergraduate settings.[11] The use of multiple formats is recommended in assessment of medical students.[12] However, assessment tools should be valid and reliable and be able to measure the different aspects of professional competencies. The present study was undertaken to evaluate MCQ and SEQ items in Basic Medical Sciences (Biochemistry) by determining item writing flaws (IWFs) of MCQs along with cognitive level of each item in both methods.

METHODS

This analytical study was carried out in the department of Biochemistry, in a newly established medical college in Karachi, Pakistan. The first batch was admitted in January 2009 while the undergraduate curriculum has been organized in six limited integrated modules. Multiple assessment methods including short essay questions, MCQs, OSPEs, and orals carrying equal weightage were used for module assessment. The college faculty developed all assessment questions. The cognitive levels of the assessment tools were analyzed using the Buckwalter’s (Buckwalter et al. 1981)[13] modification of the Bloom’s taxonomy (Bloom 1956).[14] Level I: Include questions which attempt to check recall of information. Level II: Include questions which attempt to test understanding and interpretation of data. Level III : Include questions which attempt to test the application of knowledge for solving a particular problem. For determining types of item writing flaws (IWFs) standard criteria given by Case and Swanson (2003), Haladyna et al., (2002) & Tarrant et al., (2008) were used and 14 commonly occurring violations of item-writing guidelines were identified.[8],[15],[16] One subject expert and three medical educationists reviewed each assessment tool. Initially the reviewers individually reviewed the assessment tool for the cognitive level and IWFs according to predefined criteria and reported their results to the principal investigator. A consensus meeting was called to reach on unanimous decision about the debatable questions.

RESULTS

A total 150 of MCQs that were administered in all six module examinations were reviewed. The cognitive level of 114 MCQs (76%) was at recall level while remaining 36 MCQs (24%) were of interpretation of data and there was no MCQ evaluating problem solving domain of knowledge (Table-I). A total of 36 SEQs were administered in all six module examinations. The cognitive level of 30 SEQs (83.33%) was assessing recall of knowledge while remaining 6 SEQs (16.67%) were assessing interpretation of data and there was no SEQ assessing problem solving domain of knowledge (Table-I).
Table-I

Distribution of cognitive levels of SEQs and MCQs in all six modules. (n= Number of SEQs, and MCQs in each module)

Module Cognitive levels of SEQs (n=6) Cognitive levels of MCQs (n=25)
Percentage86.11%13.89%0%76%24%0%
C1C2C3C1C2C3
M16002410
M24201870
M36002050
M45102050
M55101690
M65101690
Total 3150114360

C1= Cognitive level 1, C2= Cognitive level 2, C3= Cognitive level 3,

A total of 150 MCQs were administered in the all six module examinations. Upon review, 69 IWFs were found and four most common IWFs were implausible distracters (30.43%), unfocused stem (27.54%), unnecessary information in the stem (24.64%), and negative stem (8.7%), respectively (Table-II).
Table-II

Frequency of types of IWFs in MCQs in all six modules

Types of IWFs M1 M2 M3 M4 M5 M6 Total %
Total 2118100705869
Absolute terms000000000
Vague terms000000000
Implausible distractors 11411132130.43
Extra details in correct option10000011.44
Unfocused stems1552241927.54
Grammatical clues000000000
Logical clues03000034.35
Word repeats00100011.44
> 1 correct answer000000000
Unnecessary information in stem2633211724.64
Lost sequences in data10000011.44
All of the above000000000
None of the above000000000
Negative stem50010068.7

DISCUSSION

The selection of an appropriate assessment method for measuring students’ performances remains a daunting task for many medical institutions in Pakistan. Attempts to change existing assessment methods have been hindered by financial constraints, lack of expertise in psychometric analysis of the examination and institutional policies. The present research found that 76% MCQs were testing the recall of isolated facts while remaining 24% MCQs were testing the skill of interpretation of data and there was not a single MCQ assessing the higher cognitive domains of application and analysis. It may be due to the fact that MCQs at recall level are easier to construct and need less time and knowledge as compare to problem solving MCQs which needs expertise and training.[8],[9] Khan and Aljarallah (2011)[17] in their study found that the percentage of MCQs testing the level III (problem solving) cognitive skills of the students was 60%, level II (interpretation of data) was 6% and level I recall of information was 28%. But in that study a total of 50 MCQs representing different disciplines of medicines like gastroenterology, cardiology, neurology, rheumatology, nephrology etc were evaluated. Constructing problem solving MCQs in basic sciences is difficult in comparison to clinical sciences. In present study, multiple factors have contributed to low cognitive level questions such as newly established medical college with inadequate faculty training, diverse background of teachers, non-existence of question bank, first examination of the college etc. Tarrant and Ware (2008) [16] found in a nursing examination that over 90% of MCQs were written at low cognitive levels, and that MCQs written at a lower cognitive level were significantly more likely to contain item-writing flaws. Jozefowicz et al. (2002)[18] evaluated the quality of in-house developed examinations at three US medical schools and found that the overall quality of the questions used was low. Several studies have confirmed that MCQs not only test the knowledge of the examinees but can also be used for measuring higher cognitive skills.[19],[20] One of the most common problems affecting MCQs quality is the presence of item writing flaws. Item-writing flaws (IWFs) are violations of these accepted item-writing guidelines which can affect student performance on MCQs, making the item either easier or more difficult.[21] The present study found 69 IWFs (46%) in total 150 MCQs in Biochemistry module exams, and four most common IWF were implausible distracters (30.43%), unfocused stem (27.54%), unnecessary information in the stem (24.64%), and negative stem (8.7%). These results are in agreement with several studies.[16],[20],[22],[23] Another study documented that 12% of their exam MCQs had item writing flaws.[17] But in that study all MCQs belonged to clinical sciences subjects, and 60% of the MCQs assessed students’ problem solving skills. In our study problem solving MCQs were zero% and 76% of the MCQs assessed recall of isolated facts. Our higher percentages IWFs can be explained in the view of Tarrant and Ware (2008) study who concluded that MCQs written at lower cognitive levels, are more likely to contain item-writing flaws.[16] Downing (2005)[21] assessed the quality of four examinations given to medical students in the United States of America, and found that 46% of MCQs contained IWFs and reported that as a consequence of these IWFs, 10–15% of students who were classified as failures would have been classified as pass if items with IWFs were removed. Distribution of cognitive levels of SEQs and MCQs in all six modules. (n= Number of SEQs, and MCQs in each module) C1= Cognitive level 1, C2= Cognitive level 2, C3= Cognitive level 3, Frequency of types of IWFs in MCQs in all six modules Results of present study showed presence of 46% of flawed items which could be mostly attributed to insignificant faculty development programs. Flawed items affect difficulty and discrimination index. Low difficulty and poor discrimination in an item favors low achievers while higher difficulty and poor discrimination negatively affected the high scorers and moreover flawed items also fail to assess the course learning objectives.[23] For reducing IWFs, and improving cognitive levels of the test items, Downing 2006, suggested the use of test blue print.[24] A blue print is simply a grid or table that maps the course objectives and content to be tested, and is an essential step in generating a valid and reliable test. Test blue print helps in accurately delineating the percentage of test questions to be allocated to the different content areas, and at different cognitive levels. Faculty should be encouraged, and trained to construct MCQs for higher order cognitive levels. Tarrant et al., (2006)[25] pointed out that by removing IWFs from MCQs does not necessarily change the cognitive domain of a question, but writing questions at higher cognitive levels inherently remove numerous IWFs. The present study suggest that there is need to improve the quality of our assessment tools because if the assessment tools measur low cognitive level, it will not only decrease the validity of the exam but also compel the students to adopt surface learning approaches which is not suitable for lifelong learning. The study analyzed results of only two modules, and students’ scores in only one subject. Moreover, difficulty and discrimination indices were not available.

CONCLUSION

The medical college should evolve policy guidelines on preparing questions of higher cognitive level for all departments and student assessment should align with the learning outcomes. There is need to review the quality of item including the content of assessment tools.

Authors Contribution:

MB: Designed the study, did all analysis and prepared the manuscript. KA, SA, NH: Helped in evaluating the quality of the assessment tools, drafting and revising the manuscript.
  14 in total

1.  The quality of in-house medical school examinations.

Authors:  Ralph F Jozefowicz; Bruce M Koeppen; Susan Case; Robert Galbraith; David Swanson; Robert H Glew
Journal:  Acad Med       Date:  2002-02       Impact factor: 6.893

2.  Construct-irrelevant variance and flawed test questions: Do multiple-choice item-writing principles make any difference?

Authors:  Stevn M Downing
Journal:  Acad Med       Date:  2002-10       Impact factor: 6.893

3.  A framework for improving the quality of multiple-choice assessments.

Authors:  Marie Tarrant; James Ware
Journal:  Nurse Educ       Date:  2012 May-Jun       Impact factor: 2.082

4.  A comparison of student performance in multiple-choice and long essay questions in the MBBS stage I physiology examination at the University of the West Indies (Mona Campus).

Authors:  Dagogo J Pepple; Lauriann E Young; Robert G Carroll
Journal:  Adv Physiol Educ       Date:  2010-06       Impact factor: 2.288

5.  The effects of violating standard item writing principles on tests and students: the consequences of using flawed test items on achievement examinations in medical education.

Authors:  Steven M Downing
Journal:  Adv Health Sci Educ Theory Pract       Date:  2005       Impact factor: 3.853

Review 6.  Assessment in medical education.

Authors:  Ronald M Epstein
Journal:  N Engl J Med       Date:  2007-01-25       Impact factor: 91.245

7.  Impact of item-writing flaws in multiple-choice questions on student achievement in high-stakes nursing assessments.

Authors:  Marie Tarrant; James Ware
Journal:  Med Educ       Date:  2008-02       Impact factor: 6.251

8.  Use of an educational taxonomy for evaluation of cognitive performance.

Authors:  J A Buckwalter; R Schumacher; J P Albright; R R Cooper
Journal:  J Med Educ       Date:  1981-02

9.  Evaluation of Modified Essay Questions (MEQ) and Multiple Choice Questions (MCQ) as a tool for Assessing the Cognitive Skills of Undergraduate Medical Students.

Authors:  Moeen-Uz-Zafar Khan; Badr Muhammad Aljarallah
Journal:  Int J Health Sci (Qassim)       Date:  2011-01

10.  Assessment of higher order cognitive skills in undergraduate education: modified essay or multiple choice questions? Research paper.

Authors:  Edward J Palmer; Peter G Devitt
Journal:  BMC Med Educ       Date:  2007-11-28       Impact factor: 2.463

View more
  13 in total

1.  Concept mapping improves academic performance in problem solving questions in biochemistry subject.

Authors:  Mukhtiar Baig; Saba Tariq; Rehana Rehman; Sobia Ali; Zohair J Gazzaz
Journal:  Pak J Med Sci       Date:  2016 Jul-Aug       Impact factor: 1.088

2.  The impact of item-writing flaws and item complexity on examination item difficulty and discrimination value.

Authors:  Bonnie R Rush; David C Rankin; Brad J White
Journal:  BMC Med Educ       Date:  2016-09-29       Impact factor: 2.463

3.  Evaluation of Cognitive levels and Item writing flaws in Medical Pharmacology Internal Assessment Examinations.

Authors:  Saba Tariq; Sundus Tariq; Sadia Maqsood; Shireen Jawed; Mukhtiar Baig
Journal:  Pak J Med Sci       Date:  2017 Jul-Aug       Impact factor: 1.088

4.  Psychometric analysis of Anatomy MCQs in Modular examination.

Authors:  Zia Ul Islam; Ambreen Usmani
Journal:  Pak J Med Sci       Date:  2017 Sep-Oct       Impact factor: 1.088

5.  Evaluating the Quality of Multiple Choice Question in Paediatric Dentistry Postgraduate Examinations.

Authors:  Mawlood Kowash; Iyad Hussein; Manal Al Halabi
Journal:  Sultan Qaboos Univ Med J       Date:  2019-09-08

6.  Approaches Towards Professional Studies and Spare-time Activities Among Preclinical and Clinical Year Medical Students.

Authors:  Tahir Jameel; Mukhtiar Baig; Zohair J Gazzaz; Jawad M Tashkandi; Nasser S Al Alhareth; Shahida A Khan; Nadeem S Butt
Journal:  Cureus       Date:  2019-06-14

7.  Perceptions of medical students in Pakistan, KSA, and the US regarding the significance of case-based learning.

Authors:  Ahmed M Khalid; Mohina Sohail; Iffat Naiyar; Hajerah Khalid; Marryam Riaz; Mukhtiar Baig
Journal:  J Taibah Univ Med Sci       Date:  2021-04-09

8.  Impact of Combined modular assessment on deep learning and personal development of medical students.

Authors:  Shabana Ali; Humaira Fayyaz Khan
Journal:  Pak J Med Sci       Date:  2016 Jan-Feb       Impact factor: 1.088

9.  Effectiveness of longitudinal faculty development programs on MCQs items writing skills: A follow-up study.

Authors:  Hamza Mohammad Abdulghani; Mohammad Irshad; Shafiul Haque; Tauseef Ahmad; Kamran Sattar; Mahmoud Salah Khalil
Journal:  PLoS One       Date:  2017-10-10       Impact factor: 3.240

10.  Knowledge of dental faculty in gulf cooperation council states of multiple-choice questions' item writing flaws.

Authors:  Mawlood Kowash; Hazza Alhobeira; Iyad Hussein; Manal Al Halabi; Saif Khan
Journal:  Med Educ Online       Date:  2020-12
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.