Literature DB >> 34309243

Answering questions in a co-created formative exam question bank improves summative exam performance, while students perceive benefits from answering, authoring, and peer discussion: A mixed methods analysis of PeerWise.

Clare Guilding1, Rachel Emma Pye2,3, Stephanie Butler4, Michael Atkinson1, Eimear Field1.   

Abstract

Multiple choice questions (MCQs) are a common form of assessment in medical schools and students seek opportunities to engage with formative assessment that reflects their summative exams. Formative assessment with feedback and active learning strategies improve student learning outcomes, but a challenge for educators, particularly those with large class sizes, is how to provide students with such opportunities without overburdening faculty. To address this, we enrolled medical students in the online learning platform PeerWise, which enables students to author and answer MCQs, rate the quality of other students' contributions as well as discuss content. A quasi-experimental mixed methods research design was used to explore PeerWise use and its impact on the learning experience and exam results of fourth year medical students who were studying courses in clinical sciences and pharmacology. Most students chose to engage with PeerWise following its introduction as a noncompulsory learning opportunity. While students perceived benefits in authoring and peer discussion, students engaged most highly with answering questions, noting that this helped them identify gaps in knowledge, test their learning and improve exam technique. Detailed analysis of the 2015 cohort (n = 444) with hierarchical regression models revealed a significant positive predictive relationship between answering PeerWise questions and exam results, even after controlling for previous academic performance, which was further confirmed with a follow-up multi-year analysis (2015-2018, n = 1693). These 4 years of quantitative data corroborated students' belief in the benefit of answering peer-authored questions for learning.
© 2021 The Authors. Pharmacology Research & Perspectives published by British Pharmacological Society and American Society for Pharmacology and Experimental Therapeutics and John Wiley & Sons Ltd.

Entities:  

Keywords:  MCQ; PeerWise; assessment for learning; collaborative learning; formative assessment; gamification; medical education; peer learning; single best answer

Mesh:

Year:  2021        PMID: 34309243      PMCID: PMC8311910          DOI: 10.1002/prp2.833

Source DB:  PubMed          Journal:  Pharmacol Res Perspect        ISSN: 2052-1707


multiple choice question Newcastle University Medicine Malaysia Single Best Answer science, engineering, technology, and maths

INTRODUCTION

Background and rationale for the educational activity

Active and collaborative learning strategies, as well as frequent formative assessment with feedback, improve student learning outcomes. , , , , Students are particularly keen to engage with formative assessment that reflects their summative exams and which they can do in their own time via e‐assessment. A challenge for educators, particularly those with large class sizes is how to provide students with such learning opportunities without overburdening faculty. Multiple choice questions (MCQs) are a common form of assessment in medical degree programs. Faced with high‐stake exams, medical students revise strategically, seeking opportunities to practice MCQs to test knowledge and recall of course content. , The Newcastle University medical (MBBS) degree uses MCQ/Single Best Answer (SBA; a form of MCQ ) exams in each year of the program but does not release past exam papers. Consequently, students seek practice exam‐style questions from faculty or online question banks. Faculty have limited capacity to produce exam items for formative learning in addition to those produced for summative exams; furthermore, online question banks often charge students and the content is not curriculum specific. To address these challenges, we enrolled medical students in the online learning platform PeerWise, which enables students to author and answer MCQs/SBAs, rate the quality of other students’ contributions as well as discuss content. It provides an opportunity for self‐ and peer assessment as well as incorporating aspects of active learning. As a student generated and moderated resource it requires minimal faculty input once established.

Pedagogical principles

Constructivist learning theory posits that active self‐construction of meaning and understanding through engagement with authentic learning tasks leads to better learning, with new knowledge built on previous knowledge. , Incorporating PeerWise into a course of study creates several active learning opportunities rooted in constructivist learning theory. Writing questions for example, engages higher order thinking processes, engages students deeply in course content, and choosing distractors raises awareness of misconceptions and plausible alternative interpretations of content. , , Students writing, explaining, and discussing questions with peers are actively engaged in constructing their own knowledge and understanding of the course learning outcomes; they thereby build a collaborative learning community which studies show can enhance student performance. Practice testing has a range of cognitive, metacognitive, and noncognitive benefits. It reinforces learning, improves long‐term retention and can be a key predictor of examination performance. , , Answering questions enables students to check their understanding of course content, and provides formative and corrective feedback which are important factors in promoting learning. PeerWise has been used in learning institutions worldwide. Studies of PeerWise use and impact come predominately from undergraduate STEM subjects (science, engineering, technology, and maths); many have shown positive correlations of overall PeerWise activity with examination results. , , , , While it is difficult to disentangle the effect of the intervention from confounding factors such as other simultaneous learning activities and the intrinsic ability of students, studies in pharmacy, political science, accountancy, and STEM courses have shown improvements in assessments results when confounds such as previous academic performance are factored into the analyses. , , , Attempts to determine which aspects of PeerWise use contributed to this positive correlation indicated that academic benefits arose from engaging across a range of activities. , , Fewer studies investigate the impact of PeerWise use in medical degree programs. Some studies showed correlations between question answering, authoring, and commenting with summative exam performance, with the strongest correlations seen for question authoring. , , However, none of these studies controlled for intervention or baseline academic differences, thus students who engaged with the PeerWise activities may represent the proactive high‐achieving students who would do well on summative assessment regardless. Other studies investigating the impact of medical or pharmacy students authoring MCQs on academic achievement, using control groups of students who did not create MCQs, found no statistically significant difference between groups. , While the impact of engaging with PeerWise on academic performance in medical students remains unclear across studies, qualitative analyses of students’ perceptions of PeerWise use, in particular question answering, were generally favorable. , , Results from qualitative studies suggest that students found PeerWise an enjoyable learning experience which improved their learning competency, engagement with the materials, and motivation to learn. , , Most interventions used PeerWise as a compulsory component of the curriculum, with engagement manipulated through inclusion of PeerWise exercises in summative assessment. , , There remain contradictory reports, however, over whether medical students find compulsory MCQ writing exercises of educational benefit. , In this paper, we explore medical students’ use of PeerWise and their perceptions of the value of PeerWise as a learning tool, and investigate the relationship between PeerWise use and assessment results, controlling for previous academic performance.

METHODOLOGY

Learning environment

The Newcastle University MBBS program is 5 years in duration and delivered on two campuses: at Newcastle upon Tyne in the United Kingdom, and at Newcastle University Medicine Malaysia (NUMed) in Johor, Malaysia. Students from the United Kingdom and Malaysia are considered one cohort; the same curriculum is delivered on both campuses, students sit the same exams and they graduate with a Newcastle University MBBS degree. In the 2015 target cohort for this project, 444 fourth year students from Newcastle University (350 from United Kingdom and 94 from Malaysia) were enrolled in PeerWise at the start of Stage 4 (September 2015). In Stages 1–2 students had studied the essentials of biomedical sciences and clinical skills, Stage 3 was clinically based learning. Semester 1 of Stage 4 was primarily campus‐based learning within two main courses: Clinical Sciences and Investigative Medicine, and Clinical Pharmacology, Therapeutics, and Prescribing (hereafter referred to as Clinical Sciences, and Clinical Pharmacology, respectively). In December 2015, students took an integrated 300‐question MCQ/SBA exam assessing the content of the semester. This exam was the only knowledge‐based exam in Stage 4, and students who failed this exam failed the year and proceeded to resit in summer. Due to the integrated nature of the assessment, it was not possible to determine performance in an individual course or subject area. Students’ PeerWise use in Semester 1 and performance in the end of semester “Stage 4 exam” was investigated for this project.

PeerWise use

Students were introduced to the Stage 4 Clinical Pharmacology and Clinical Sciences PeerWise courses through an email to the year group which provided information on how to use the site, the pedagogic rationale and potential benefits, together with detailed written instruction on how to write MCQ/SBA questions. Students were asked to write a question each, include explanations for the correct answer and distractors, tag their question with a topic for question sorting; and they were able to answer, comment on and rate other contributors’ questions. Engagement with PeerWise was not a compulsory part of the curriculum; PeerWise was not included within set teaching and learning activities and participation was independent of summative Stage 4 assessment. Thus, students’ use of PeerWise was self‐directed and interaction with the site was for formative rather than summative assessment and feedback. The PeerWise courses were established partway through Semester 1 for the 2014 cohort. At the start of the 2015 academic year, the Clinical Pharmacology course contained 65 questions and the Clinical Sciences course contained 123 questions written by the previous cohort and available to the target cohort of students to answer.

Research design

This study aimed to explore PeerWise use and its impact on the learning experience and exam results of fourth year medical students. The research questions were: How do students engage with authoring, answering, and commenting on formative assessment questions in PeerWise? Does student engagement with PeerWise predict performance in Stage 4 exams, when taking Stage 3 performance into account, and are particular elements of PeerWise, such as authoring questions, driving this relationship? What are students’ perceptions of the benefits and challenges associated with engaging in authoring, answering, and commenting on formative assessment questions in PeerWise? A quasi‐experimental, convergent parallel mixed methods research design was used. Data collection were concurrent: at the start of Semester 2 following completion of the Stage 4 exam, quantitative data were generated automatically and downloaded directly from the PeerWise system, and a qualitative survey was distributed to students assessing their perceptions and use of PeerWise. Qualitative and quantitative data were analyzed separately, and the strands mixed during interpretation (see Section 4). This mixed methods approach was chosen for (a) triangulation––qualitative survey statements of perceived benefits of PeerWise use compared with quantitative analysis of impact on exam performance, and (b) expansion––extending the breadth of the enquiry by using the qualitative and quantitative methods for the different research questions. The research was approved by the School of Medical Education Research Management Group and ethical approval gained from the Faculty of Medical Sciences Ethics Committee at Newcastle University (reference #1608/2015). All participants were advised in writing of the aims of the research, that their consent to participate was voluntary, that the data may be used for publication and they could withdraw their consent and contributions at any time.

Data collection and analysis

Student engagement measures

Measures of student engagement were automatically downloaded from the PeerWise website. To measure question complexity of student‐generated MCQs, two researchers categorized a random sample of 100 questions written by the 2015 cohort, as follows: knowledge, comprehension, application, analysis, evaluation, and creation, based on Anderson and Krathwohl's revised version of Bloom's taxonomy. Questions were categorized independently, and any discrepancies were discussed until agreed.

Quantitative data analysis

Five measures are available from PeerWise. The authoring, answering, and rating scores incorporate the number of contributions together with how a student's questions are rated by the cohort and how often the cohort subsequently agreed with their ratings. The reputation score is used as a marker for overall engagement with PeerWise and is a combination of the three component scores which favors students participating in all areas. In effect, reputation is a measure of a student's engagement with PeerWise, the timeliness of such engagement (with scores increasing as a result of subsequent student behavior, encouraging early engagement) and how similar their responses were to their peers’ later responses. A later addition, the answer score increases each time a student answers a question correctly with a small number of points deducted for incorrect answers, somewhat independent from others’ performance. We therefore focused on reputation as a measure of overall engagement, and answer as a measure of knowledge recall. Our analysis required both Stage 3 and Stage 4 exam scores, therefore we excluded data from students who did not sit both exams. Exam scores were matched to PeerWise usernames and anonymized. All analysis was then carried out on anonymized data. Quantitative data analysis was calculated using IBM SPSS (version 26).

Survey

A survey was developed by the researchers and students with previous PeerWise experience to explore students’ participation with PeerWise, and perceived hindrances and benefits to engagement. The survey was hosted on Bristol Online Survey and underwent content and face validity checks by the members of the team who developed and reviewed it before distribution to the 2015 cohort of Stage 4 students via email. The survey asked the students whether they had authored, answered, or commented on questions, and dependent on their response asked for free text comments on what prevented them from engaging, or about any perceived benefits. A final question asked the students which elements of PeerWise they believed most supported their learning and why. Free‐text comments in the survey were analyzed following the established reflexive thematic analysis approach detailed by Braun and Clarke. Survey responses were downloaded directly into Microsoft Excel and read by researchers to identify “patterns of meaning and issues of potential interest” (p. 15). Researchers independently categorized the responses into broad preliminary codes, then through discussion combined these into the final codes for data analysis. Codes were analyzed, and related codes combined to generate subthemes and overarching themes. Subthemes were quantified to facilitate examination of the magnitude of responses.

RESULTS

Student engagement with PeerWise

Engagement of the 2015 target cohort with the various elements of PeerWise is shown in Table 1. We evaluated the cognitive levels the student‐authored questions tested. Thirteen percent of the questions were categorized as testing knowledge, 28% comprehension, 18% application, 18% analysis, 12% evaluation, and 11% creation. Cognitive levels tested in the Clinical Pharmacology course were skewed more toward the lower end of Blooms taxonomy than questions in the Clinical Sciences course (see Figure 1).
TABLE 1

Engagement of students with question answering, authoring, commenting, and rating in PeerWise during Stage 4 Semester 1 (September–December 2015). The “number engaged” column is the number of students who participated in each element of PeerWise

PeerWise courseClinical sciencesClinical pharmacology
# Students (%) a 359 (81%)310 (70%)
Activity measure (number)Number engagedMeanMedianMaxTotalNumber engagedMeanMedianMaxTotal
Questions answered35916213734359,920308999918431,124
Questions authored470.5049203280.3023104
Question comments971.4048535750.8021260
Questions rated2721076233839,7492426648.518020,753

The number in brackets shows this value as a % of the total class of 444.

Only includes students who activated their PeerWise account.

FIGURE 1

Categorization of 50 clinical sciences and 50 clinical pharmacology questions according to Blooms taxonomy. The Y axis shows the number of questions within each category

Engagement of students with question answering, authoring, commenting, and rating in PeerWise during Stage 4 Semester 1 (September–December 2015). The “number engaged” column is the number of students who participated in each element of PeerWise The number in brackets shows this value as a % of the total class of 444. Only includes students who activated their PeerWise account. Categorization of 50 clinical sciences and 50 clinical pharmacology questions according to Blooms taxonomy. The Y axis shows the number of questions within each category

Quantitative PeerWise analysis

Stage 4 scores pre‐ and post‐PeerWise

To determine if Stage 4 exam performance was higher after the introduction of PeerWise, we used a repeated‐measures ANOVA of Stage 3 and 4 results across two cohorts prior to PeerWise introduction (2012, n = 322 and 2013, n = 370), and following PeerWise introduction (2015, n = 444 and 2016, n = 431). We did not use 2014 cohort data as PeerWise was established part way through the semester. We first determined that each students’ Stage 3 and Stage 4 results significantly correlated within each cohort, using Pearson's correlation (2013 r = .701; 2014: r = .727; 2015: r = .734; 2016: r = .686; all values of p < .001). Stage 3 scores remained relatively consistent across cohorts with more variation across Stage 4 (see Figure 2). There was a significant main effect of Stage (F(1, 1563) = 7603.211, p < .001), with Stage 4 scores significantly lower than Stage 3, and a significant interaction between Stage and Cohort (F(3, 1563) = 86.8, p < .001). Using Stage 4 minus Stage 3 difference scores to control for cohort differences in the baseline Stage 3 measures, we ran a one‐way ANOVA using a priori contrasts to compare pre‐ and post‐PeerWise performance. There was a significant difference, with a smaller score drop post‐PeerWise than pre‐PeerWise (t(1563) = 10.46, p < .001), implying a significant benefit of PeerWise on Stage 4 exam performance.
FIGURE 2

Mean Stage 3 and Stage 4 knowledge exam scores over four cohorts: two cohorts before PeerWise introduction (2012, 2013) and two after (2015, 2016). Y axis shows the mean exam score for each cohort, error bars are standard error of the mean

Mean Stage 3 and Stage 4 knowledge exam scores over four cohorts: two cohorts before PeerWise introduction (2012, 2013) and two after (2015, 2016). Y axis shows the mean exam score for each cohort, error bars are standard error of the mean

2015 cohort analysis

A mean of Clinical Pharmacology and Clinical Science scores for the five PeerWise measures reputation, answering, authoring, rating, and answer was calculated for the 2015 cohort. A linear regression was performed with Stage 4 score as the predicted value, and Stage 3 score entered as the first variable in Model 1, with reputation and answer entered into Model 2. Stage 3 score explained 53.7% of the variance (Model 1: F(1, 442) = 515.245, p < .001). While Model 2 was significant and predicted 55.3% of the variance, with a significant R 2 change (F(2, 440) = 8.988, p < .001), the only significant variable was mean answer (t = 3.751, p < .001), with mean reputation not a significant predictor (t = 0.407, p = .684). This pattern is also seen when analyzing each PeerWise course separately. As Authoring has been specifically proposed as a mechanism for better learning, we also regressed mean answering, authoring, and rating as separate variables predicting Stage 4, after entering Stage 3 into the model. The model was significant (F(4, 439) = 136.548), explaining 55% of the variance. Only mean answering was a significant predictor (t = 3.641, p < .001), with mean rating (t = −1.226, p = .221) and mean authoring (t = −0.77, p = .441) nonsignificant predictors.

Quartile splits

Following previous analyses, , we aimed to identify if certain groups of students benefitted from PeerWise use during their course, so repeated the analyses above for each quartile, removing Stage 3 from Model 1. The students scoring higher at Stage 3 (Q1 and Q2) showed the same pattern as the whole cohort (Q1: F(2, 119) = 3.566, p = .031; Q2: F(2, 101) = 5.269, p = .007), with a significant prediction from mean answer but not mean reputation of Stage 4 results. The models with students who scored lower at Stage 3 (Q3 and Q4) were not significant (Q3: F(2, 117) = 1.636, p = .199; Q4: F(2, 95) = 1.871, p = .16), with no significant prediction from either mean reputation or mean answer of Stage 4 results.

Multi‐year analysis

As more cohorts use PeerWise, students’ scores on reputation accumulate to reflect the quality of the questions they asked during their enrolment on a particular course. We were interested to explore whether this cumulative indicator of “reputation” was related to Stage 4 performance. Across four cohorts (2015–2018; n = 1695), an hierarchical regression was conducted with Stage 4 results as predicted variable. Stage 3 results were entered first into the model (F(1, 1694) = 1298.12, p < .001) then mean reputation and mean answer (F(3, 1691) = 466.06, p < .001). Correlations between the variables were all highly significant, reflecting the large sample size (see Table 2). However, the regression analysis indicated that while mean answer was a significant predictor of Stage 4 results, controlling for Stage 3 (t = 7.475, p < .001), mean reputation was not (t = −0.846, p = .397). This corroborates the 2015 analysis, showing over four cohorts, a significant positive relationship between answering PeerWise questions and exam results, even after controlling for previous academic performance.
TABLE 2

Pearson's correlations between Stage 4 and Stage 3 exam scores, and PeerWise mean values of reputation and answering across the two target courses for 2015–2018 cohorts

Stage 4Stage 3Mean reputation
Stage 3.659***
Mean reputation.077** .083***
Mean answer.205*** .106*** .284***

p <.01.

p <.001.

Pearson's correlations between Stage 4 and Stage 3 exam scores, and PeerWise mean values of reputation and answering across the two target courses for 2015–2018 cohorts p <.01. p <.001.

Student perception of the benefits and challenges of PeerWise for learning

There were 174 responses (38.2% of the 2015 cohort) to the survey which assessed students’ use of PeerWise and perceived benefits and challenges to participation. The results from the survey question that asked students to rate which elements of PeerWise they found most beneficial are shown in Table 3. Thematic analysis of free text data identified responses in three overarching themes: answering, authoring, and commenting on questions. Themes were composed of subthemes related to what students found beneficial and challenging in these areas (see Table 4). Below, we describe the themes and subthemes, presenting illustrative data extracts.
TABLE 3

Student ratings of perceived relative benefit of different aspects of PeerWise for learning on a 5‐point Likert scale, 1 being the least benefit, 5 being the most

N MeanSEM
Answering questions1674.400.07
Reading explanations of answers1674.370.06
Writing explanations1084.050.1
Writing questions1023.770.12
Evaluating quality of questions1522.970.1
Commenting on questions1282.840.1

Abbreviations: N, number of student responses; SEM, standard error of the mean.

TABLE 4

Themes and subthemes identified from thematic analysis of student survey responses

ThemeSubtheme (n)
Answering

Benefits from answering

Identifying gaps in knowledge (43)

Testing and consolidation of learning (41)

Exam technique and practice (29)

Explanations improve knowledge/understanding (15)

Wide range of curriculum relevant questions (13)

Benchmarking against peers (9)

Novel, active revision method (9)

Challenges of answering

Questions too niche/difficult (15)

Inconsistent question quality (13)

Questions didn't reflect exam (4)

Difficulty using the site (2)

Authoring

Benefits of authoring

Stimulated learning through in‐depth research into topic (13)

Consolidation of knowledge/understanding (12)

Insight into how exam questions are composed (9)

Ensured thorough understanding of topic for good quality, error free question (8)

Writing distractors and explanations helps identify confounding information (4)

Identification of gaps on knowledge (4)

Challenges of authoring

Lack of time (41)

Concern about question writing ability (32)

Concern over negative peer feedback (7)

Used other revision method (6)

Difficult/time‐consuming to write questions (5)

Bank already full of questions (5)

Unsure of benefit of authoring (4)

Commenting

Benefits of commenting

For clarification of question/answers (24)

Correction of incorrect knowledge/understanding (13)

To help and encourage peers (11)

Generates peer discussion (10)

Explaining answers reinforces learning (7)

Challenges of commenting

Comments already covered what student would have raised (8)

Did not feel the need (6)

Sought clarification in course materials (3)

Concerned comment may be incorrect (2)

Abbreviation: n, number of comments within each subtheme.

Student ratings of perceived relative benefit of different aspects of PeerWise for learning on a 5‐point Likert scale, 1 being the least benefit, 5 being the most Abbreviations: N, number of student responses; SEM, standard error of the mean. Themes and subthemes identified from thematic analysis of student survey responses Benefits from answering Identifying gaps in knowledge (43) Testing and consolidation of learning (41) Exam technique and practice (29) Explanations improve knowledge/understanding (15) Wide range of curriculum relevant questions (13) Benchmarking against peers (9) Novel, active revision method (9) Challenges of answering Questions too niche/difficult (15) Inconsistent question quality (13) Questions didn't reflect exam (4) Difficulty using the site (2) Benefits of authoring Stimulated learning through in‐depth research into topic (13) Consolidation of knowledge/understanding (12) Insight into how exam questions are composed (9) Ensured thorough understanding of topic for good quality, error free question (8) Writing distractors and explanations helps identify confounding information (4) Identification of gaps on knowledge (4) Challenges of authoring Lack of time (41) Concern about question writing ability (32) Concern over negative peer feedback (7) Used other revision method (6) Difficult/time‐consuming to write questions (5) Bank already full of questions (5) Unsure of benefit of authoring (4) Benefits of commenting For clarification of question/answers (24) Correction of incorrect knowledge/understanding (13) To help and encourage peers (11) Generates peer discussion (10) Explaining answers reinforces learning (7) Challenges of commenting Comments already covered what student would have raised (8) Did not feel the need (6) Sought clarification in course materials (3) Concerned comment may be incorrect (2) Abbreviation: n, number of comments within each subtheme.

Answering questions theme

93% of survey respondents had answered questions on PeerWise and seven subthemes were identified as perceived benefits of answering (see Table 4). Students commented that answering questions “was hugely beneficial as it showed me the gaps in my knowledge as well as allowing me to see what my peers were revising. It was also very reassuring when I started getting questions right.” The testing of knowledge and getting answers right increased students’ confidence “it made me feel confident that I was on the right track with my learning.” Identification of gaps in knowledge prompted students to do more reading around the subject, and in many cases, completion of questions taught them new material: “Answering questions was a confidence booster if I got the answer correct, and getting them wrong gave me areas to revise again. Also getting them wrong made me learn it, I often found in the real exam I knew the answer because of PeerWise questions.” Many students appreciated the provision of practice exam‐style questions, in part because “they were questions—something that are rarely provided to us.” Students felt answering questions got them into the mind‐set of the exam, helping them think about how questions and the distractors are constructed. Comments included that it was helpful “seeing the possible other options made me more aware of confounders and ‘second best’ answers to be aware of.” Another subtheme identified as a benefit of answering PeerWise questions was the wide range of curriculum relevant questions, as illustrated by these quotes: “the questions were designed by us and focused specifically on our learning outcomes,” “the fact that the questions are written by people on my course most supported my learning, often other online exam tools are very generic and not focused on my outcomes” and “there were many examples of many different questions and the breadth and depth of questions allowed for questioning of how well you knew the subject.” Students valued the explanations given for the correct/incorrect answers and used these to improve their knowledge and understanding, commenting “the section explaining the answer and other answers was helpful. It helped you remember and understand the condition,” and “reading the explanations was often helpful (if they were provided) as they were written in an accessible way, especially if they explained why certain answers were incorrect.” The final beneficial subthemes identified included: benchmarking progress against peers, and that answering questions was a novel, active revision method, highlighted by the comments “I thought it was beneficial in allowing me to see what my peers felt I should know for my exams. Medicine has such width and depth of learning sometimes it can be hard to gauge the level I am aiming for with my revision. This tool allowed me to see what my peers considered reasonable exam questions,” and “I was able to see if I was catching up with the rest of the cohort or if I was lacking behind and more work was required of me. It also helps to reactivate things that I had been reading and apply it through some of the brilliant and creatively written questions”. Four subthemes were identified around challenges or hindrances to engagement with answering PeerWise questions. Difficult or poor‐quality questions had a negative emotional impact on some students who commented “some questions were demoralizing as they focused in too much on minute details of the curriculum.” Quality of questions led one student to query whether the site could be trusted for revision “I felt that some of the answers to the questions were debatable. Also, I didn't know if I could trust the answers that were provided, and I preferred not to take the risk with my revision.”

Authoring of questions theme

32% of respondents said they had authored a question, and a fifth of these said the process had consolidated their knowledge or understanding. A further fifth of the respondents said writing questions stimulated them to do further research which increased their learning: “it made me read to ensure I thoroughly understood the topic before creating a question.” Part of the drive to do in‐depth research was motivated by the pressure to ensure the question was high quality and error free: “I felt responsible to make a water tight question so that it couldn't be disputed which encouraged me to revise more.” Writing questions helped some respondents identify gaps in their knowledge: “it takes a greater understanding to write a question than just to answer one, especially in the MCQ format. By trying to write questions, I found gaps in my knowledge that I wouldn't have recognised otherwise.” Students who authored on PeerWise commented that the process gave them helpful insights into how exam questions are composed and the areas of the curriculum that lend themselves to testing by MCQ/SBA: “it gave me an insight into how question authors might be thinking, particularly with regards to phrasing; this definitely influenced my answers in the exam.” Writing distractors (incorrect answers), explanations and getting feedback from peers were highlighted by students as beneficial: “the most helpful parts were thinking of incorrect but plausible answers (good for forming differentials) and especially writing an explanation. I feel like the topics I wrote a question on were much easier to answer in the exam.” 68% of respondents said they had not authored a question with a quarter of all respondents citing a lack of time, many of them stating they used PeerWise for last minute revision. Some said they would have written a question had they started using it earlier, and they could see the potential benefits of authoring: “Time constraints because I started using Peerwise too late… If I had accessed Peerwise earlier, I might have because authoring a question would be very beneficial as it stimulates our thinking and links up all our thoughts.” A fifth of respondents said they did not author because they were concerned about their ability to write a good quality question, this was often related with concern about what their peers would think and a lack of confidence, for example “I didn't feel that I would be able to produce a question that was detailed enough to be of value to others.” Some students were explicit that that they did not author because of concern over negative peer feedback as illustrated by these quotes “Literally peer pressure! The fear of making an error in the question/answer.” The time and difficulty in authoring questions inhibited some students, who chose to stick with established learning techniques: “I wrote separate, more basic, Anki questions and couldn't think of any good ‘exam’ style questions,” “it would take a while to write a question (and check it rigorously to make sure it is correct/good) and time is limited whilst I am revising.” Others reflected that they did not author because the PeerWise courses already had many questions or they were unsure of the benefit: “often the topics I would have been confident writing questions in were already very well represented with questions,” “a lack of belief that it would be personally beneficial for me.”

Commenting on questions theme

41% of survey respondents said they had engaged in commenting on PeerWise questions. The most common reason given for commenting was to clarify elements of the question or answers: “It helped clarify what the correct answer was and why when it was not clear—it was possible to post in explanations and text from lectures etc to show where the answer could be found etc,” and “Commenting allowed me to voice any concerns about either my thought process or the structure of the question. The author was then able to tell me their perspective so I understood the question better and was therefore more able to learn from it.” Commenting enabled students to correct errors in others’ knowledge or understanding, or have their own errors corrected, students stating they commented “to correct a poor question—helped me understand the subject” and “the author corrected me on a point I had misunderstood.” The flagging of an error enabled authors to correct their question: “Reading the comments was useful, I pointed out some information had been missed out which was shortly corrected.” A fifth of students who commented said they did so to help and encourage their peers, through praise for questions, use of humor to lighten the mood while studying or through sharing of their knowledge. Students noted they commented “to motivate authors for contributing questions and express gratitude for their efforts,” and “the only comment I made was a jovial, light‐hearted jest… Perhaps it offered others some much needed humour enabling them to continue on their slogs for a longer period of time!” The commenting function of PeerWise provided a platform for peer discussion, and through discussion and explanation of answers students were able to better understand the topics, as illustrated by these quotes: “I discussed some questions I felt inaccurately represented the topic or were misleading and this helped both my and the question author's understanding,” and “it was most useful when people were willing to engage in constructive debate about why an answer was appropriate or not appropriate. Commenting required me to articulate my knowledge clearly and that process tested whether I actually understood what I was talking about!” 59% of the respondents said they had not used the commenting function on PeerWise. The most common reason given for not responding was that the comments already covered what student would have raised, “I didn't have much to say as quite a few people had already commented with what I might have said.” Students said they did not feel the need to comment, sometimes clarifying that this was because the questions were good enough “I never disagreed with the questions, or felt I had anything else to contribute to them.” A couple of students noted they would rather seek clarification from external sources “if I was unsure I like to refer back to my own notes, textbooks, internet sources.”

DISCUSSION

This study found that most students engaged with PeerWise following its introduction as a noncompulsory learning opportunity in Stage 4 (Table 1). While students perceived benefits in authoring and peer discussion, students engaged most highly with answering questions, noting that this helped them identify gaps in knowledge, test their learning, and improve exam technique (Table 4). Quantitative analysis corroborated the perception that answering questions was beneficial to learning. Analysis of the Stage 3 and 4 knowledge exams showed a significantly smaller decrease in exam score following the introduction of PeerWise (Figure 2). Detailed analysis of the 2015 cohort indicated that, after controlling for previous academic performance, PeerWise use significantly predicted Stage 4 exam result underpinned by answering questions but not by authoring, peer discussion, or overall engagement. Multi‐year analysis of four cohorts supported the 2015 results, showing a sustained significant positive predictive relationship between answering PeerWise questions and exam performance, after controlling for pre‐PeerWise exam results. Most published studies have integrated PeerWise as a taught, compulsory component of the curriculum with engagement contributing to summative assessment to encourage student participation. , , Lower participation has been found when no marks were awarded. , We introduced PeerWise as a non‐taught, noncompulsory component of the curriculum, and in contrast, observed high participation, at least with regards to question answering. Studies that investigated the cognitive level assessed by student‐created MCQs found the majority of items in STEM courses assessed at the lower levels of Blooms taxonomy. , , , , Students in this study have been at university longer than those in most STEM courses, having three years’ previous experience passing MCQ/SBA exams. Thus, the questions created by our students ranged more widely in complexity, including analysis and evaluation questions (Figure 1). In previous studies, students frequently noted answering questions as the biggest benefit of using PeerWise. , , The perceived benefits of answering such as improved course knowledge, improved exam taking skills, and increased confidence are similar across this study and others. , , , Frequent formative assessment with explanatory feedback has been shown to be valuable in promoting learning. , Explanatory feedback aids the learning and retention of information, , and explanations of MCQs which offered solutions to the questions were highlighted in this study as an important learning tool. Practice testing of MCQs has direct and indirect effects on student learning. , Indirectly, frequent tests stimulate more frequent studying thus more time‐on‐task, while directly, the retrieval of information during tests results in deeper learning and greater knowledge retention than more passive forms of learning such as reading notes. , Our data on exam performance following PeerWise engagement support the perceived benefits of answering questions, our hierarchical regression models consistently showing a significant relationship between question answering and exam performance. Unlike other studies which indicated that the benefits of PeerWise arose from students overall level of activity, , we did not find a significant relationship between reputation score and exam performance. Interestingly, when we looked at the impact of PeerWise use in ability quartiles, our data indicated that students in the higher two quartiles of ability were significantly benefited, while the lower quartiles were not. Previous studies have shown a range of results regarding the benefits according to ability quartiles. Hardy et al. showed year 1–2 students in the lower intermediate ability levels on STEM programs may benefit most, while other studies showed students in the highest and lowest quartiles in year 1 may benefit most. , The overall ability in our cohort of fourth year medical students likely differs from those of early year STEM students, making comparisons of PeerWise impact in ability quartiles across studies difficult. Only a minority of students engaged with question authoring and peer discussion compared to question answering (see Table 1). This result is consistent with previous studies of medical students which showed that even when engagement is linked to summative assessment, or when authoring and peer discussion exercises are included as a taught component of the curriculum, answering questions remained the most popular element. , , In this study, PeerWise was used over one semester, with only 12 weeks of intense teaching before the Stage 4 exam. Key reasons cited by the students for not authoring was lack of time and doubt over ability. Authoring MCQs is challenging and time‐consuming. It is the challenging nature of the process which educators perceive drives deep learning and understanding; question writing is suggested to employ higher levels of comprehension than answering. , , Indeed, students who had authored in our study believed this stimulated in‐depth learning and understanding. Similar results were found in other studies, which showed that question authoring activities drove desirable learning behaviors such as the integration of varying types and sources of information. , The feasibility and acceptability of question writing as a study technique for medical students has, however, been questioned. , A compulsory question writing exercise for a pathology course within a medical degree was recently removed due to low student acceptance. Researchers have suggested that the question writing may not be an efficient learning strategy within time pressured programs, , and our study did not show a significant association between question writing and improved exam performance. Introducing PeerWise as a noncompulsory learning opportunity allowed our students to engage with the elements they perceived most effective for their learning, and may improve acceptance in other interventions. A concern over question writing ability and a fear of negative peer feedback further inhibited students from authoring. Walsh et al. identified “PeerWise trolling” in their intervention and suggested that introductory sessions define clear rules for online interactions. In our correspondence with students it was made clear that while the site was anonymous for students, faculty could identify participants and modify interactions. Since launch, 2143 fourth year students have used the site, yet only one interaction was flagged as inappropriate and removed. Providing face‐to‐face training in question writing should improve students’ confidence in authoring, would result in a larger question bank, and has been requested by other UK medical students. Medical students who received in‐person training in question writing reported that it gave them insights into how to strategically approach MCQs which they believed improved their test taking aptitude ; these are perceived benefits also noted by authors in this study. Yet, not all students are motivated and have the self‐regulation skills to write MCQs, plus given the time constraints of the semester a noncompulsory training session may be more acceptable to the cohort.

Practical implications

Our study shows significant benefits of answering student‐authored MCQs on exam performance. Building a bank of MCQs for formative assessment is time‐consuming and expensive. By engaging students in question authoring, large banks of curriculum relevant questions can be generated easily. Beyond the initial account setups and distribution of instructions and usernames each year, there has been minimal faculty input. Even with minimal input, between 2014 and 2020, Stage 4 students authored 1648 questions for the Clinical Pharmacology and Clinical Sciences courses, which were answered 842,298 times, generating 5402 comments. Arguably, introducing taught question writing sessions would generate a greater number of questions, potentially of higher quality ; Gooi and Sommerfeld provide guidance for introducing such question writing sessions into curricula.

Strengths and limitations

Limitations of this research include that it is limited to a single course within one institution reducing generalizability of findings, although the large sample size may allow for a level of relatability to other contexts. The study did not incorporate a control group due to ethical considerations of allowing only part of the cohort access to an intervention that may have been beneficial, and the response rate to the survey was <40% so different themes could have arisen if all students had responded. The strengths of this research include the large volume of PeerWise data that was available for analysis, collected from students over 4 years. A significant proportion of students were accessing and using PeerWise during the study adding to the credibility of findings. The hierarchical regression analysis allowed for control of previous academic ability in assessment of the impact of PeerWise on summative exam performance, while the mixed methods approach allowed for a rich exploration and triangulation of data that allowed us to gain a better understanding of the barriers and facilitators to engagement with education tools by undergraduate students.

CONCLUSION

As educators, we look to develop students as lifelong independent learners, so that they become increasingly self‐regulating and self‐directed as they mature into competent graduates. This study suggests that PeerWise provides such opportunities by encouraging learners to work independently and collaboratively to construct their own learning. Despite its introduction as a noncompulsory learning activity, most students used PeerWise and described a range of perceived benefits, which alongside the positive correlation of PeerWise use and summative exam performance support its introduction within programs and can be used as evidence to encourage future participation within medical education contexts.

DISCLOSURE

The authors have no conflict of interest to declare.

ETHICS STATEMENT

The research was approved by the School of Medical Education Research Management Group and ethical approval gained from the Faculty of Medical Sciences Ethics Committee at Newcastle University (reference #1608/2015). All participants were advised in writing of the aims of the research, that their consent to participate was voluntary, that the data may be used for publication and they could withdraw their consent and contributions at any time.

AUTHOR CONTRIBUTION

CG, REP, and EF designed the study; CG and EF acquired the data; REP performed the quantitative statistical analysis; CG, SB, and MA performed the quantitative data analysis. CG and REP wrote the paper; SB and MA critically revised the paper. All the authors have revised and approved the final version of the manuscript.
  22 in total

1.  Constructing multiple choice questions as a method for learning.

Authors:  Edward Palmer; Peter Devitt
Journal:  Ann Acad Med Singapore       Date:  2006-09       Impact factor: 2.473

2.  Assessment in medical education.

Authors:  Christopher M Buttery
Journal:  N Engl J Med       Date:  2007-05-17       Impact factor: 91.245

3.  The critical importance of retrieval for learning.

Authors:  Jeffrey D Karpicke; Henry L Roediger
Journal:  Science       Date:  2008-02-15       Impact factor: 47.728

4.  Using mixed methods research in medical education: basic guidelines for researchers.

Authors:  Karen E Schifferdecker; Virginia A Reed
Journal:  Med Educ       Date:  2009-07       Impact factor: 6.251

5.  Medical school 2.0: How we developed a student-generated question bank using small group learning.

Authors:  Adrian C C Gooi; Connor S Sommerfeld
Journal:  Med Teach       Date:  2014-10-13       Impact factor: 3.650

6.  Crowdsourcing for assessment items to support adaptive learning.

Authors:  Sean Tackett; Mark Raymond; Rishi Desai; Steven A Haist; Amy Morales; Shiv Gaglani; Stephen G Clyman
Journal:  Med Teach       Date:  2018-08-10       Impact factor: 3.650

7.  Peer generation of multiple-choice questions: student engagement and experiences.

Authors:  Susan M Rhind; Graham W Pettigrew
Journal:  J Vet Med Educ       Date:  2012       Impact factor: 1.027

8.  Improving large class performance and engagement through student-generated question banks.

Authors:  Dale Hancock; Nicole Hare; Paul Denny; Gareth Denyer
Journal:  Biochem Mol Biol Educ       Date:  2018-03-12       Impact factor: 1.160

9.  Medical students create multiple-choice questions for learning in pathology education: a pilot study.

Authors:  Rebecca Grainger; Wei Dai; Emma Osborne; Diane Kenwright
Journal:  BMC Med Educ       Date:  2018-08-22       Impact factor: 2.463

10.  Formative student-authored question bank: perceptions, question quality and association with summative performance.

Authors:  Jason L Walsh; Benjamin H L Harris; Paul Denny; Phil Smith
Journal:  Postgrad Med J       Date:  2017-09-02       Impact factor: 2.401

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.