Literature DB >> 27672617

Interactive learning in oral and maxillofacial radiology.

Aruna Ramesh1, Rumpa Ganguly1.   

Abstract

PURPOSE: The use of electronic tools in teaching is growing rapidly in all fields, and there are many options to choose from. We present one such platform, Learning Catalytics™ (LC) (Pearson, New York, NY, USA), which we utilized in our oral and maxillofacial radiology course for second-year dental students.
MATERIALS AND METHODS: The aim of our study was to assess the correlation between students' performance on course exams and self-assessment LC quizzes. The performance of 354 predoctoral dental students from 2 consecutive classes on the course exams and LC quizzes was assessed to identify correlations using the Spearman rank correlation test. The first class was given in-class LC quizzes that were graded for accuracy. The second class was given out-of-class quizzes that were treated as online self-assessment exercises. The grading in the self-assessment exercises was for participation only and not accuracy. All quizzes were scheduled 1-2 weeks before the course examinations.
RESULTS: A positive but weak correlation was found between the overall quiz scores and exam scores when the two classes were combined (P<0.0001). A positive but weak correlation was likewise found between students' performance on exams and on in-class LC quizzes (class of 2016) (P<0.0001) as well as on exams and online LC quizzes (class of 2017) (P<0.0001).
CONCLUSION: It is not just the introduction of technological tools that impacts learning, but also their use in enabling an interactive learning environment. The LC platform provides an excellent technological tool for enhancing learning by improving bidirectional communication in a learning environment.

Entities:  

Keywords:  Assessments, Educational; Computer-Assisted Instruction; Education, Dental; Radiology

Year:  2016        PMID: 27672617      PMCID: PMC5035726          DOI: 10.5624/isd.2016.46.3.211

Source DB:  PubMed          Journal:  Imaging Sci Dent        ISSN: 2233-7822


Introduction

Evolving methods of student learning necessitate the evolution of teaching methods. Computers have become an integral part of both our personal and professional lives, and an obvious result of this is the introduction of computers and personal electronic devices into the domains of teaching and learning. The body of literature discussing technology in education is expanding, including studies on the use of technology in education in the health fields.12 Currently, much attention is paid to guided learning, rather than simply presenting information in the traditional 'sage on the stage' method. Knowles3 based his proposed learning theory on the premise that learning is most likely to occur when learners are self-directed and motivated, with a perceived need to acquire information relevant to their own area of experience. A blended learning approach is a combination of traditional lecture-format classroom teaching with use of self-directed online and technology-enabled learning tools. The use of technologically based learning tools enables interaction and informal self-testing in a non-threatening way.456 Tools that allow the learner to actively engage in their own learning, such as quizzes with feedback, have the advantage of being perceived as low-stakes assessments, rather than a high-stakes assessments that use pass/fail criteria.4789 Classroom response systems are a set of technology tools that can enhance student participation and learning while enabling educators to ascertain what their students are learning.10 With the use of classroom response systems such as clickers or other web-based systems, continuous feedback can be obtained on student understanding. One such technology is Learning Catalytics™ (LC) (Pearson, New York, NY, USA), which is a 'bring-your-own-device' student engagement, assessment, and classroom intelligence system. LC allows faculty to obtain real-time or out-of-class responses to open-ended or critical thinking questions, in order to determine which areas require further discussion and clarification. The comprehensive analytics help faculty better understand student performance. At our institution, we adopted LC in the pre-doctoral oral and maxillofacial radiology (OMFR) curriculum for dental students in academic year 2013-2014. We have utilized this technology tool for the last two years in this course for the DMD classes of 2016 and 2017. The OMFR course is taught in the second year of the dental school curriculum. It spans 8 months and the class meets once a week for an hour. The course is divided into 3 sections. The first section addresses concepts in radiation physics, biology and protection, and radiographic techniques. The second section addresses the identification of normal radiographic anatomy, the detection of dental caries, and periodontal pathology. The third section introduces the radiographic interpretation of pathology, culminating in the generation of differential diagnoses for various entities spanning developmental anomalies and benign, malignant, systemic, and dysplastic lesions. Since this is the first time the dental students are introduced to radiography and radiographic interpretation, it is typically a challenging experience. The introduction of LC in this class had the primary goal of evaluating students' understanding of the material on an ongoing basis and providing feedback or clarification as needed. Since the inclusion of any additional teaching tool necessitates curriculum revision with time allotment for these exercises, this study aimed to evaluate the effectiveness of LC as a specific learning enhancement tool in the field of the diagnostic sciences. The objective of this study was to assess the correlation between students' performance on course exams and LC quizzes. The hypothesis was that the students' performance in course exams and LC quizzes in the second-year predoctoral OMFR course would not vary.

Materials and Methods

This study was a retrospective analysis of the OMFR grades obtained by the dental classes (DMD) of 2016 and 2017 in their second year of study. The OMFR course taught in academic years 2013-2014 and 2014-2015 utilized the LC program to administer quizzes during the course. In academic year 2013-2014 (class of 2016), inclass LC quizzes were graded for the accuracy of the answers and were administered 1-2 weeks before the course exams. The material included in the quiz was the same as the topics included in the following course exam. The course contained a total of 3 examinations and 3 quizzes. In academic year 2014-2015 (class of 2017), the LC quizzes were given outside of class as online self-assessment exercises. The grading was based on participation only and not accuracy. The online LC quizzes were likewise made available 1-2 weeks before the course examinations, and the course contained 3 examinations and 5 quizzes. The maximum of the allotted points for each quiz for both classes was 5. The number of students in the classes of 2016 and 2017 who completed all quizzes and exams was 171 and 183, respectively. This study was a retrospective analysis of the correlation of the students' grades obtained on examinations 1-3 and the LC quiz grades in the OMFR course in both academic years (classes of 2016 and 2017). Institutional Review Board approval was obtained for this study. A convenience sample of 354 subjects was obtained from the DMD students in the classes of 2016 and 2017. Descriptive statistics (means, standard deviations, and medians for categorical variables) are presented in Tables 1 and 2. Statistical significance for the association between LC quiz scores and exam scores was assessed using the Spearman rank correlation test, as the data were not normally distributed. The assumption of normality was examined graphically and by the Shapiro-Wilk test. P-values <.05 were considered to indicate statistical significance. Stata version 13.1 (StataCorp LP, College Station, TX, USA) was used for the analysis. The class of 2016 had 3 LC quizzes and the class of 2017 had 5 LC quizzes. For the purposes of our analysis, in light of the topics covered and the timing of the quizzes, the average scores of quizzes 2 and 3 and the average scores of quizzes 4 and 5 were used as single items for the class of 2017.
Table 1

Descriptive statistics comparing the classes of 2016 and 2017 (mean±standard deviation)

Table 2

Descriptive statistics comparing the classes of 2016 and 2017 (median and interquartile range)

Results

A positive but weak correlation was found between overall quiz scores and exam scores when the 2 classes were combined; that is, higher quiz scores were associated with higher exam scores. A positive but weak correlation was also found between students' performance on examinations and in-class LC quizzes (class of 2016), as well as between students' performance on examinations and online LC quizzes (class of 2017). No correlation was found between the exam and quiz scores for individual exams and quizzes; the correlations were only found for aggregated exam and quiz scores. These data are presented in Table 3.
Table 3

Correlation between the exams and the Learning Catalytics™ quizzes using the Spearman rank correlation

A comparison of the average exam score per student per class with the average quiz score found no significant differences in the median average exam score or median average quiz score between the classes of 2016 and 2017 (Table 4).
Table 4

Comparison of average exam and quiz scores for the classes of 2016 and 2017

Discussion

The integration of technological tools into everyday life has increased the availability and use of such tools in education as well. The second-year predoctoral students in the OMFR course used their own smartphones or devices to log in to the LC program and access the quiz modules. The primary goal of introducing the quizzes before each exam in the course was to encourage earlier preparation for exams and to provide an opportunity for self-assessment and clarifying the subject matter with the instructors, leading to better understanding of the material. Our objective in this study was to evaluate the correlation between students' performance on exams and the LC quizzes. We found a direct overall correlation between quiz and exam performance, indicating that LC quizzes served as predictors of performance on the course exams. Our results are in agreement with those of Nayak and Erinjeri,11 who compared the use of the Audience Response System (ARS) and end-of-course examinations and reported that no significant differences were observed between the ARS results and those of end-of-course examinations (2.38 versus 2.04, P<.165). These quizzes also served as feedback to the students, as they could perform self-assessment and improve their understanding. Feedback has been listed as a major influence in learning. Figure 1 presents a screenshot of an LC window showing a question, the student's response and its accuracy, the number of students who responded to this question, and an explanation of why option A is the correct answer. Hattie and Timperley12 defined feedback as a "consequence" of performance and as information provided by an agent (e.g., a teacher, a book, a parent, oneself, or experience) regarding aspects of one's performance or understanding. To quote Kulhavy's13 1977 formulation, "Feedback has no effect in a vacuum; to be powerful in its effect, there must be a learning context to which feedback is addressed. It is but part of the teaching process and is that which happens second, after a student has responded to initial instruction, when information is provided regarding some aspect(s) of the student's task performance." The feedback model that we adopted for both the classes of 2016 and 2017 was aligned with this philosophy. For the class of 2016, immediate in-class feedback was directed at clarifying concepts relating to the incorrect responses and reinforcing the correct responses.
Fig. 1

The Learning Catalytics™ module shows the student view with the correct response and explanation.

For the class of 2017, students were allowed to respond to each question multiple times as they completed the quiz outside of class; students received feedback on each attempt, but only the last response was recorded. This allowed for continuous feedback as they completed the quiz. For both years, the students had access to the LC quiz module throughout the course, including during preparation for exams. Since the LC quiz grades and exam grades were directly correlated, albeit weakly, we are unsure whether the feedback had a direct impact on exam performance. The evaluation we performed in this study did not allow us to determine whether students' performance improved between the quiz and exam, but only found a correlation between these 2 parameters. However, the students indicated that the quizzes and feedback they received were excellent learning experiences. Similar to this study, Nayak and Erinjeri11 found that when comparing the use of ARS and end-of-course examinations to recognize students' mastery of course material, students reported no significant difference (1.77 versus 1.85, P< .712). In the same study, in terms of how to study for a course, students reported significantly increased insights from the use of an ARS rather than hearing verbal responses to questions asked by the lecturer (2.38 versus 3.27, P<.002).11 The LC platform can be used to obtain real-time feedback during class. In our study, for the class of 2016, the quizzes were scheduled and administered in class, followed by a brief feedback session. In contrast, for the class of 2017, the quizzes were completed online, the quizzes could be attempted more than once, and the students had a week to complete them. Despite these differences in the LC quiz formats; no statistically significant difference was found in the correlation of the quizzes with exam scores between the 2 classes. It is interesting that the class of 2016 was graded on accuracy while the class of 2017 was only graded for participation. Despite this seemingly major discrepancy, no significant difference was observed in the quiz grades for these two classes. The accuracy of the quiz grades for class of 2017 was obtained and compared to the accuracy of the class of 2016 for the purposes of the study. The longer time frame to complete the quiz, ability to take the quiz at their own convenience, and the potentially lower anxiety related to grading only for participation and not accuracy did not seem to have made any difference in how the LC quiz grades related to exam grades. This leads to the question of whether it is necessary to grade self-assessment quizzes, especially if immediate feedback is provided; however, how seriously the quizzes would be taken by students if they are not graded is a question worth considering. The reason for our switch to online quizzes for the class of 2017 was that management of the LC module in class took away from classroom instruction time allotted traditional lectures. This has been observed as an issue by other authors.14 Since no difference was found in student performance and learning between the in-class and online quiz models, class time may be structured to include more interactive or case-based education if quizzes do not take up class time. ARSs have been in use for almost half a century and various surveys have found medical residents enjoy ARS lectures more than other didactic techniques. Many authors have shown that ARSs not only boost attentiveness, but also improve the short and long-term retention of lecture material. While most studies relating to ARSs and other technological tools have focused on medical education, 15161718 our study focused on dental education, and specifically oral and maxillofacial radiology. Our students also had positive comments regarding the use of LC modules in our course. Meckfessel et al.19 assessed student perspectives on the use of e-learning tools in dental radiology and received favorable feedback. The authors also reported a positive correlation between the 'e-program' and improved test scores. Their e-program provided an independent interactive learning tool for the students. The LC tool that we used was more of an extension of what the lecturers taught in class and allowed us to continuously monitor how the students were learning, which enabled continuous modifications in our teaching. One of the biggest advantages of using any audience response technology is the anonymity it provides students. It allows for a bidirectional flow of questions between the instructor and students. Students may feel uncomfortable exposing their lack of knowledge when answering or asking questions in the presence of their peers. LC is useful in this regard, as the instructor can send out spontaneous, on-the-fly questions to the class and vice versa, fairly quickly. In a large class, such as in our course with more than 170 students, reading and responding to all questions posted in class did require a substantial time investment outside of class for the instructors. Nonetheless, it did open up a channel of effective communication between the class as a whole and the instructors. In the final course evaluations, both the classes of 2016 and 2017 had positive feedback about the use of LC in the radiology course. With only 3 exams in the course, the additional quizzes in LC were perceived as aids to understanding the material. Many students in the class of 2017 commented positively on the online LC quizzes. Two of the assessment statements in the evaluation survey used for all didactic courses at the school that specifically addressed learning in the course are: "This course effectively promoted learning" and "The course material enhanced previous learning." The students used a 6-point scale in the evaluation (1, strongly agree; 2, agree; 3, neutral; 4, disagree; 5, strongly disagree; 6, not applicable). The mean scores of the class of 2016 for these 2 statements were 1.59 and 1.64 respectively, while the mean scores of the class of 2017 for these 2 statements were 1.56 and 1.57 respectively. The difference in the format of the LC quizzes did not impact student-perceived learning in the course. Of course, the LC modules were a small part of the entire course and these student assessments dealt with the entire course structure. Despite the large class size, every attempt was made by the course instructors to make the course as interactive as possible by encouraging class participation and discussion. These student evaluations refer to this interactive course model, which also includes the LC modules. We are in agreement with Richardson16 in our strong belief that learning cannot be affected just by the introduction of a technological tool alone, but the impact that technological tools can have in enabling an interactive learning environment is impressive. The LC platform provides an excellent technological tool for enhancing learning by improving bidirectional communication in a learning environment.
  11 in total

1.  Audience response system: effect on learning in family medicine residents.

Authors:  T Eric Schackow; Milton Chavez; Lauren Loya; Michael Friedman
Journal:  Fam Med       Date:  2004 Jul-Aug       Impact factor: 1.756

2.  Collaborative testing as a learning strategy in nursing education.

Authors:  Sheryl S Sandahl
Journal:  Nurs Educ Perspect       Date:  2010 May-Jun

3.  Introduction of e-learning in dental radiology reveals significantly improved results in final examination.

Authors:  Sandra Meckfessel; Constantin Stühmer; Kai-Hendrik Bormann; Thomas Kupka; Marianne Behrends; Herbert Matthies; Bernhard Vaske; Meike Stiesch; Nils-Claudius Gellrich; Martin Rücker
Journal:  J Craniomaxillofac Surg       Date:  2010-05-07       Impact factor: 2.078

4.  Challenges in healthcare education: a correlational study of outcomes using two learning techniques.

Authors:  Karen Ray; Brenda Berger
Journal:  J Nurses Staff Dev       Date:  2010 Mar-Apr

5.  Use of an audience response system to augment interactive learning.

Authors:  Robyn Latessa; David Mouw
Journal:  Fam Med       Date:  2005-01       Impact factor: 1.756

6.  Audience response systems in medical student education benefit learners and presenters.

Authors:  Lina Nayak; Joseph P Erinjeri
Journal:  Acad Radiol       Date:  2008-03       Impact factor: 3.173

7.  Demonstrating the effectiveness of an online, computer-based learning module for arterial blood gas analysis.

Authors:  Jessica Schneiderman; Susan Corbridge; Julie Johnson Zerwic
Journal:  Clin Nurse Spec       Date:  2009 May-Jun       Impact factor: 1.067

Review 8.  Creating a computer-assisted learning module for the non-expert nephrology nurse.

Authors:  Hazel A Dennison
Journal:  Nephrol Nurs J       Date:  2011 Jan-Feb       Impact factor: 0.959

Review 9.  Interactive, technology-enhanced self-regulated learning tools in healthcare education: a literature review.

Authors:  Julia Petty
Journal:  Nurse Educ Today       Date:  2012-07-18       Impact factor: 3.442

10.  Audience response techniques for 21st century radiology education.

Authors:  Michael L Richardson
Journal:  Acad Radiol       Date:  2014-05-13       Impact factor: 3.173

View more
  4 in total

Review 1.  Artificial intelligence for precision education in radiology.

Authors:  Michael Tran Duong; Andreas M Rauschecker; Jeffrey D Rudie; Po-Hao Chen; Tessa S Cook; R Nick Bryan; Suyash Mohan
Journal:  Br J Radiol       Date:  2019-07-26       Impact factor: 3.039

2.  An systematic review of e-learning outcomes in undergraduate dental radiology curricula-levels of learning and implications for researchers and curriculum planners.

Authors:  Michael G Botelho; Kalpana R Agrawal; Michael M Bornstein
Journal:  Dentomaxillofac Radiol       Date:  2018-08-01       Impact factor: 2.419

3.  The development of a learning management system for dental radiology education: A technical report.

Authors:  Hee-Jin Chang; Khanthaly Symkhampha; Kyung-Hoe Huh; Won-Jin Yi; Min-Suk Heo; Sam-Sun Lee; Soon-Chul Choi
Journal:  Imaging Sci Dent       Date:  2017-03-21

4.  Effectiveness of Problem-Based Learning versus Traditional Teaching Methods in Improving Acquisition of Radiographic Interpretation Skills among Dental Students-A Systematic Review and Meta-Analysis.

Authors:  Alexander Maniangat Luke; Simy Mathew; Sam Thomas Kuriadom; Jeny Mary George; Mohmed Isaqali Karobari; Anand Marya; Ajinkya Mansing Pawar
Journal:  Biomed Res Int       Date:  2021-09-24       Impact factor: 3.411

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.