Literature DB >> 29943399

Academic outcomes of flipped classroom learning: a meta-analysis.

Kuo-Su Chen1,2,3, Lynn Monrouxe2,3, Yi-Hsuan Lu1,2,3, Chang-Chyi Jenq2,3,4, Yeu-Jhy Chang2,3,5, Yu-Che Chang2,3,6, Pony Yee-Chee Chai7.   

Abstract

CONTEXT: The flipped classroom (FC), reversing lecture and homework elements of a course, is popular in medical education. The FC uses technology-enhanced pre-class learning to transmit knowledge, incorporating in-class interaction to enhance higher cognitive learning. However, the FC model is expensive and research on its effectiveness remains inconclusive. The aim of this study was to compare the efficacy of the FC model over traditional lecture-based (LB) learning by meta-analysis.
METHODS: We systematically searched MEDLINE, PubMed, ERIC, CINAHL, EMBASE, reference lists and Association for Medical Education in Europe (AMEE) conference books. Controlled trials comparing academic outcomes between the FC and LB approaches in higher education were considered eligible. The main findings were pooled using a random-effects model when appropriate.
RESULTS: Forty-six studies (9026 participants) were included, comprising four randomised controlled trials (RCTs), 19 quasi-experimental studies and 23 cohort studies. Study populations were health science (n = 32) and non health science (n = 14) students. The risk of bias was high (36/37 articles). Meta-analyses revealed that the FC had significantly better outcomes than the LB method in examination scores (post-intervention and pre-post change) and course grades, but not in objective structured clinical examination scores. Subgroup analyses showed the advantage of the FC was not observed in RCTs, non-USA countries, nursing and other health science disciplines and earlier publication years (2013 and 2014). Cumulative analysis and meta-regression suggested a tendency for progressively better outcomes by year. Outcome assessments rarely focused on behaviour change.
CONCLUSIONS: The FC method is associated with greater academic achievement than the LB approach for higher-level learning outcomes, which has become more obvious in recent years. However, results should be interpreted with caution because of the high methodological diversity, statistical heterogeneity and risk of bias in the studies used. Future studies should have high methodological rigour, a standardised FC format and utilise assessment tools evaluating higher cognitive learning and behaviour change to further examine differences between FC and LB learning.
© 2018 The Authors. Medical Education published by Association for the Study of Medical Education and John Wiley & Sons Ltd.

Entities:  

Year:  2018        PMID: 29943399      PMCID: PMC6120558          DOI: 10.1111/medu.13616

Source DB:  PubMed          Journal:  Med Educ        ISSN: 0308-0110            Impact factor:   6.251


Introduction

The flipped classroom (FC) is a hybrid approach, combining online learning and face‐to‐face classroom activities. In this pedagogical model, students engage in content learning before class, thereby maximising in‐class time for active learning.1 The FC utilises technology for pre‐class learning, with face‐to‐face classrooms becoming interactive learning activities. This methodology restructures and reorders traditional lecture‐based (LB) approaches by moving students, rather than teachers, to the centre of learning.2, 3, 4 Such active learning should improve outcomes as learners practise with, engage with and apply their pre‐class learning.5 Although modern versions of the FC appeared over 10 years ago,6 and despite its popularity in education generally and medical education specifically, we lack firm conclusions regarding its efficacy.7, 8 The FC approach is underpinned by active learning theory.9 Active learning has been defined as ‘any instructional method that engages students in the learning process’.10 An active learning approach is one in which students undertake meaningful learning activities, cognitively engaging in those activities. Previous research suggests that students' understanding and performance are improved via active learning.10, 11, 12, 13 A meta‐analysis confirmed that active learning increased student performance and decreased failure rate in undergraduates.14 Thus, as FC is a form of active learning,4, 15 it has been associated with better learning outcomes,16 especially in higher‐order thinking.17, 18 Indeed, research suggests that an FC approach increases motivation, satisfaction, performance and class attendance rate.4, 19, 20 However, research results are generally inconsistent,2, 19, 21, 22, 23, 24, 25, 26, 27, 28, 29 even when the same outcome measure is considered.7, 8 A number of systematic reviews have been undertaken in the area of health care professionals' education that highlight this issue of inconsistency. For example, in a systematic review of the effectiveness of the FC in medical education, Chen et al.7 found nine articles reporting randomised or non‐randomised controlled studies and a further 37 reporting studies with ‘other’ methods or descriptive accounts (e.g. action research, pre–post designs and commentaries). Although students' perceptions of the FC process were typically positive across all nine of the controlled studies, Chen et al., 7 reported mixed results in terms of changes in knowledge, and skills. For example, regarding outcomes using multiple‐choice questions (MCQs), some studies reported positive findings for the FC over the LB approach,30, 31 whereas others found no difference.22, 32 Furthermore, Chen et al. found a varying direction and magnitude of the effect sizes and confidence intervals (CIs) across studies.7 A more recent systematic review by Hughes and Lyons8 also found mixed results across 11 studies when considering MCQ outcomes: four studies each demonstrated an advantage and a disadvantage of the FC approach over the LB approach and a further three studies showed mixed results. Systematic reviews have also been undertaken in the area of nurse education, reporting similar ambiguous results. For example, Betihavas et al.33 identified only five studies for inclusion in their systematic review. They found that although the FC condition tended to produce better outcomes immediately post‐event26, 34 alongside positive course evaluations,35, 36 when it came to performance in the final examination, any advantage was lost.26 Furthermore, not all studies reported an advantage for the FC,2 and not all course evaluations were positive.34 Betihavas et al.33 concluded that there was a lack of evidence for the efficacy of the FC, claiming that much could be learned by examining how ‘other health disciplines, such as pharmacy, have contributed extensively to implementing flipped classroom models’. Furthermore, in a narrative review of 13 articles reporting on the FC method in nurse education, Presti20 concluded that ‘few studies exist on the flipped classroom in nursing education, and only one study statistically validates its value in improving examination scores’. Finally, a recent integrative review also found 13 empirical studies reporting FC outcomes in nursing education (with some overlap in studies reported by Betihavas et al33 and Presti20).27 Unsurprisingly, it too concluded that results from FC studies were mixed. As we can see from these literature reviews, we are left with a rather muddy picture regarding the efficacy of FC over LB learning. Given the importance of the FC, it seems pertinent that we employ research methods to further understand whether the FC method works and in what circumstances. Data aggregation by meta‐analysis37, 38, 39 can settle controversies from conflicting studies, allowing greater power and precision to estimate the effectiveness of relative interventions, and can generate new hypotheses. In a meta‐analysis, the cause of inconsistency between studies can be explored through subgroup analyses and meta‐regression. Thus, a meta‐analysis may help to answer questions that individual studies cannot. Following this notion, two meta‐analyses have very recently been published by researchers from China, drawing on studies examining the FC method in nursing.40, 41 Hu et al.40 focused purely on randomised controlled trials (RCTs) of the FC method with baccalaureate nursing students in China between 2015 and 2017. They found 11 RCTs matching their inclusion and exclusion criteria, which measured two broad outcomes: theoretical knowledge, and skills. In terms of theoretical knowledge, only one of the nine studies reporting this outcome demonstrated no difference between the FC and LB approaches. Results of meta‐analyses showed a significant difference between the FC and lecture, favouring the FC approach. For the outcome of students’ skills scores, only five studies were included in the meta‐analysis, with only one study demonstrating no difference between learning conditions. Again, the pooled effect size found an overall significant difference favouring the FC over the LB condition. Tan et al.41 also undertook a meta‐analysis of studies reporting the relative effectiveness of the FC in Chinese nursing education (2014–2016). They included 29 articles in their review, measuring theoretical examination scores (n = 16), skills examination scores (n = 16) and measures of self‐learning via questionnaires (n = 15). Briefly, they found significant differences in favour of the FC approach for all three outcome measures.41 Despite the number of recent reviews examining the relative efficacy of the FC over the LB approach, a gap in the literature remains: all of the reviews were limited to a single discipline (e.g. nursing education33, 40, 41 and medical education7, 8), with a small number of articles generally favouring the FC approach (often between five and 11, with only one including 29), mostly being narrative syntheses with only two meta‐analyses.40, 41 These two meta‐analyses considered data from a single country (China), only pooling RCTs from Chinese nursing education published in Chinese journals. The results of such meta‐analyses from this very restricted data source may be difficult to generalise. Given that the FC pedagogical model has been adopted across a wide range of disciplines, data pooling across multiple disciplines may help to increase statistical power and identify differences in the effects across different domains. Thus, a more inclusive approach, not restricted to a single discipline or single country, can open up the possibility of multiple subgroup analyses enabling an understanding of the nuances around the relative efficacy of the FC approach in comparison with LB learning. We therefore developed a study using a meta‐analysis technique that was not limited to a single discipline or country, to answer the following research questions (RQs). RQ1 Is the FC learning approach associated with students’ enhanced knowledge and behaviour more than the LB learning approach in higher education? RQ2 Does this differ by context (e.g. research method, publication year, publication forum, student group and learning environment)?

Methods

The meta‐analysis was undertaken following the PRISMA (preferred reporting items for systematic reviews and meta‐analyses) statement (see Box 1 for definitions of terms used in a meta‐analysis).42

Meta‐analysis

A statistical analysis that combines the results of multiple quantitative information from related studies and produces results that summarise a whole body of research.

Begg's test

An indicator of publication bias. Begg et al. proposed testing the interdependence of variance and effect size using Kendall's method. This bias indicator makes fewer assumptions than that of Egger et al. but it is insensitive to many types of bias to which Egger's test is sensitive.

Cochrane Q

A classical measure of heterogeneity in meta‐analysis. Q is distributed as a chi‐squared statistic with n − 1 degrees of freedom. For the Q‐statistic test, a ‘Cochrane Q‐value > degree of freedom’ or a ‘p‐value of < 0.10’ suggests the presence of heterogeneity.

Egger's test

Another indicator of publication bias. Egger et al. suggested a test to examine asymmetry in funnel plots. This is a test for the Y‐intercept = 0 from a linear regression of the normalised effect estimate (i.e., the estimate divided by its standard error) against precision (the reciprocal of the standard error of that estimate).

Funnel plot

A graph that provides a visual check of the existence of publication bias. In the absence of publication bias, it assumes that studies with high precision will be plotted near the average, and studies with low precision will be spread evenly on both sides of the average, creating a roughly funnel‐shaped distribution. Asymmetry in funnel plots may indicate the presence of publication bias in meta‐analysis.

Heterogeneity

Heterogeneity in meta‐analysis refers to the variation or inconsistency in the study outcomes between studies.

I‐squared (I 2)

Another measure of heterogeneity in meta‐analysis. The I² statistic is the percentage of variation across studies attributable to heterogeneity. The formula of I² = 100% × (Q–df)/Q. An I 2 of < 25% represents low heterogeneity, 25–50% moderate heterogeneity and > 50% high heterogeneity.

Meta‐regression

A statistical method that can be used in meta‐analysis to examine the impact of moderator variables on study effect size using regression‐based techniques.

PRISMA statement

A guideline for the reporting of systematic reviews and meta‐analyses. The PRISMA statement comprises 27 items and a four‐phase flow diagram.

Publication bias

Occurs when the published results depend not just on the quality of the research but also on the hypothesis tested, and the significance and direction of effects detected. It suggests that results not supporting the hypotheses of researchers often go unpublished, leading to a bias in published research.

Random‐effects model

The process of meta‐analysis is undertaken by either a fixed‐effect or a random‐effects statistical model. A fixed‐effect meta‐analysis assumes all studies are estimating the same (fixed) treatment effect, whereas a random‐effects meta‐analysis permits differences in the treatment effect across studies. In a random‐effects model, the observed heterogeneity is attributed to two sources: (i) between‐study heterogeneity in true effects, and (ii) within‐study sampling error.

Eligibility criteria

The criteria for study eligibility are listed in Table 1 and include: one comparator study examining the FC against the LB approach; undergraduate students or higher; assessment of learning, behaviour change and impact of behaviour change, rather than student satisfaction,43 and outcomes quantitatively measured. Duplicate reports were excluded.
Table 1

Inclusion and exclusion criteria

Inclusion criteriaExclusion criteria
Two groups, controlled studies (randomised or non‐randomised, interventional or observational)Reviews, editorials, qualitative studies or single‐arm before–after studies
Research question meets the following PICO (Problem, Intervention, Comparison, Outcome) criteria: (P) higher education (any discipline, any level) (I) flipped classroom learning in any format (C) traditional lecture‐based learning (O) academic performance (any Kirkpatrick level 2 to level 4 learning outcome)27 Outcome only reporting satisfaction or perception survey
K‐12 education
Control group not traditional lecture‐based learning
Duplicate reporting (studies from same author or institution)
Quantitative outcomes data availableSelf‐rated or self‐assessed academic outcomes
Post‐intervention assessment, change in pre–post intervention assessmentPre‐intervention assessment or assessment during the process of intervention, such as quiz or homework
Any language
Inclusion and exclusion criteria

Data sources

The search took place in June 2016. Data sources included electronic databases, reference lists, conference abstracts and dissertations. Electronic databases included MEDLINE, PubMed, the Cumulative Index to Nursing and Allied Health Literature (CINAHL), the Education Resources Information Center (ERIC) and Excerpta Medica Database (EMBASE). Related publications that appeared on the web page during the electronic search (e.g. ‘similar articles’ appearing on the screen in PubMed, or ‘other users also viewed this article’ appearing in ScienceDirect) were enrolled if eligible. Reference lists of relevant manuscripts were searched for additional literature. Theses or conference abstracts appearing in the reference lists or the electronic searches were included if eligible. The Association for Medical Education in Europe (AMEE) conference books (2007–2016) were searched for unpublished studies.

Search strategy

The key search terms used in the electronic database searches were ‘flipped classroom’, ‘flipped education’, ‘flipped learning’, ‘reverse classroom’, ‘backward classroom’, ‘inverted classroom’ and ‘inverse classroom’.

Study selection

Study selection was based on the a priori set of criteria for inclusion and exclusion: for articles that clearly met the inclusion criteria, or ‘possibly’ met them, full‐text publications were retrieved for the formal review. Each study was independently assessed by at least two authors (Y‐CC, C‐CJ and Y‐JC). All disagreements between them were resolved by discussion or third‐party adjudication (K‐SC).

Assessment of methodological quality

The Effective Public Health Practice Project (EPHPP) Quality Assessment Tool44 was used to appraise methodological quality as it is suitable for both interventional (RCT, quasi‐experimental) and observational (cohort, historical control and case–control) studies. With this tool, each study was rated as strong, moderate or weak according to the following components: selection bias; study design; confounding factors; study blinding; data collection; withdrawals; and dropouts. A global rating was then allocated using the standard system: strong (no weak ratings); moderate (one weak rating); or weak (two or more weak ratings). Risk of bias was only assessed in journal articles and theses. Conference abstracts were not appraised for risk of bias because of the limited information available. Risk of bias was assessed by two authors independently (Y‐HL and PY‐CC). Disagreements between them were again resolved by discussion or third‐party adjudication (K‐SC).

Data synthesis

Data were independently extracted by two reviewers (Y‐HL and K‐SC) using a standardised form. The extracted information included publication type and year, design, demographics of participants, intervention and control, and type of outcomes measured. The cognitive levels of learning outcomes measured were categorised according to the Kirkpatrick model43 and Bloom's taxonomy.45 Given that multiple effect measures are usually reported in educational research,46, 47, 48, 49 several methods have been proposed to manage multiple effect measures in meta‐analysis.50 In this work, we pooled data separately based on the type of effect measure. When several similar effect measures were reported in the same study, we selected the one that best represented the particular type of outcome required for the meta‐analysis.50 Data pooling was undertaken by combining the standardised mean difference (SMD) using a random‐effects model.51 If a study reported several datasets for the same type of effect measure, the averaged mean and pooled variance was used for data pooling.52 For studies that did not report standard deviations (SDs) for their estimates, or had other missing data, we contacted the author by e‐mail first. If no response from authors was forthcoming, we synthesised the SD from the CIs or exact p‐values, interquartile ranges and minimum–maximum by the principle of estimation method as previously reported.53, 54, 55 We eliminated studies from data pooling if the required information for meta‐analysis (such as participant number, mean and SD) could not be obtained. Heterogeneity of effect sizes across studies was evaluated using the Q‐statistic and quantified by I‐squared values.56 Subgroup analyses were performed to explore the impacts of different contexts on the outcome measures. Cumulative analyses and meta‐regression were undertaken to examine the relationship between publication year or sample size and effect size. Publication bias was evaluated through visual inspection of funnel plots and Egger's regression test. Meta‐analyses were carried out using stata 11.2 software (StataCorp LP, College Station, TX, USA).

Results

Study search and selection

The process is summarised in a PRISMA42 flow diagram (Fig. 1). The electronic search resulted in 1682 potentially relevant citations. An additional 24 studies were added through another source. These 1706 retrieved records were pooled into a single database in EndNote (Clarivate Analytics, Philadelphia, PA, USA). After removing 428 duplicates, 1278 citations were subject to title and abstract screening and 114 studies were subject to a full‐text review. Finally, 46 studies with a total of 9026 participants were included. The characteristics of these studies are summarised in Table S1.
Figure 1

PRISMA flow chart of the study selection process

PRISMA flow chart of the study selection process

Scope of the included studies

Of the 46 included studies, 36 were journal articles, nine were conference abstracts and one was a thesis (Table S1). In terms of design, they included four RCTs, 19 quasi‐experimental studies and 23 cohort studies. Thirty‐seven studies originated from the USA and nine from other countries. There were 32 health science and 14 non‐health science education studies. Among the 32 health science education studies, 14 involved medical education, nine pharmacy education, three nursing education and six were from other health science education disciplines.

Outcomes measured

Multiple effect measures were reported. We classified the effect measures into three categories: examination scores (n = 41); course grades (n = 9); and objective structured clinical examinations (OSCEs) (n = 2). The examination scores were the results of MCQ tests (n = 20) and written examinations or essays (n = 3), a combination of different formats (n = 5) or were unclear (n = 13). Of the 41 studies reporting examination scores, seven were before–after studies and provided pre–post score changes. All examination score outcomes evaluated knowledge change according to Kirkpatrick's level 2 measurement (learning).43 The level of cognitive learning evaluated in these tests comprised only lower cognitive levels in one study, ranging from low to high (levels 1–6 in Bloom's taxonomy45) in 16 studies and were not described in 24 studies. Multiple assessments, such as homework assessments, attendance rates, quizzes, presentations, and mid‐term and final examination scores, were combined for the calculation of a grade score. The formula for the course grade calculation differed between studies.

Quality assessment of the included studies

Methodological quality assessment (36 journal articles, one thesis) showed that 36 studies exhibited a high risk of bias and one study had a moderate risk of bias (see Table S2 for details). The weakest indicator was blinding. Risks were also high in the data collection, confounders and withdrawal domains. Poor reporting quality was an important cause of a high risk of bias. Many quality indicators were not clearly described or were absent. Frequently missing information included: study design; eligible population; participant and dropout populations; important confounders between groups at baseline, and the reliability/validity of outcome assessment tools. This resulted in many studies being rated with a high risk of bias.

Main and subgroup meta‐analyses

We now report the main and subgroup meta‐analyses to address our two related RQs. Is the FC learning approach associated with better academic performance than the LB learning approach in higher education? Does this differ by context (e.g. research method, publication year, publication forum, student group and learning environment)? Figures S1–S9 are available to visually support our interpretation and we will refer to them throughout. Following Valentine et al., 57 we note that it is possible to undertake a meta‐analysis with as few as n = 2 studies.

Meta‐analyses by outcome measures

Data pooling of all 46 studies was undertaken according to the type of outcome measure reported. Significantly higher examination scores, pre–post examination score improvements and course grades for the FC approach were identified (Table 2, Fig. 2). However, no significant effect was noted when the outcome measure was an OSCE score. Statistical heterogeneity was very high for all types of outcome measured.
Table 2

Meta‐analyses according to discipline and academic outcomes

Type of outcome measuredTest for heterogeneityTest for effect
I 2, % Q‐statisticp‐valuePooled effect size (95% CI)p‐value
All studies (n = 46)a
Examination scores (n = 41)90.5422.42<0.0010.468 (0.307, 0.629)<0.001
Pre–post‐test change (n = 7)82.634.54<0.0010.454 (0.139, 0.769)0.005
Course grade (n = 9)76.033.33<0.0010.354 (0.157, 0.552)<0.001
OSCE (n = 2)98.774.18<0.0013.116 (−2.219, 8.451)0.252
Health science studies (n = 32)a
Examination scores (n = 28)86.9205.67<0.0010.451 (0.277, 0.625)<0.001
Pre–post‐test change (n = 5)82.322.58<0.0010.603 (0.264, 0.943)<0.001
Course grade (n = 5)49.47.910.0950.440 (0.245, 0.636)<0.001
OSCE (n = 2)98.774.18<0.0013.116 (−2.219, 8.451)0.252
Non‐health science studies (n = 14)a
Examination scores (n = 13)94.4216.19<0.0010.490 (0.143, 0.837)0.006
Pre–post‐test change (n = 2)33.21.500.221−0.022 (−0.425, 0.381)0.916
Course grade (n = 4)81.816.500.0010.274 (−0.080, 0.629)0.130
Medical education studies (n = 14)a
Examination scores (n = 14)80.566.59<0.0010.53 (0.31, 0.74)<0.001
Pre–post‐test change (n = 4)80.215.150.0020.71 (0.31, 1.10)<0.001
OSCE (n = 1)5.86 (4.77, 6.95)
Pharmacy education studies (n = 9)a
Examination scores (n = 7)92.580.39<0.0010.53 (0.12, 0.93)0.011
Course grade (n = 3)35.93.120.2100.53 (0.35, 0.71)<0.001
Nursing education studies (n = 3)a
Examination scores (n = 2)79.04.760.0290.20 (−0.33, 0.73)0.468
Course grade (n = 1)0.05 (−0.38, 0.49)
OSCE (n = 1)0.41 (−0.17, 1.00)
Other disciplines in health science (n = 6)a
Examination scores (n = 5)89.437.74<0.0010.20 (−0.34, 0.74)0.470
Pre–post‐test change (n = 1)0.26 (0.03, 0.49)
Course grade (n = 1)0.29 (−0.21, 0.78)

Some studies in this classification had more than one outcome measure. CI = confidence interval.

Figure 2

Meta‐analyses of all 46 studies by outcome measure. CI = confidence interval; RCT = randomised controlled trial; SMD = standardised mean difference

Meta‐analyses according to discipline and academic outcomes Some studies in this classification had more than one outcome measure. CI = confidence interval. Meta‐analyses of all 46 studies by outcome measure. CI = confidence interval; RCT = randomised controlled trial; SMD = standardised mean difference Data pooling for the 32 health science education studies showed similar results to the main analysis: the FC condition had higher examination scores, higher course grades and better pre‐ to post‐test changes, but similar OSCE scores (Table 2, Fig. S1). Data pooling for the 14 studies in non‐health science education showed a better outcome for the FC condition in examination scores, but not in the pre‐ to post‐test scores or course grades (Table 2, Fig. S2). Detailed analyses regarding the outcomes reported for medical education, pharmacy education, nursing education and other disciplines in health sciences education separately by outcome measure were undertaken (see Table 2 and Figs S2–S6 for all details). Briefly, examination scores in the FC condition were significantly better in medical and pharmacy education but not in nursing and other health disciplines, and course grades were significantly better in pharmacy education. We were not able to undertake an analysis for some outcomes because of a lack of studies.

Subgroup analyses by context

A subgroup analysis was undertaken for the outcome of examination scores. This showed an advantage of the FC condition in medical, pharmacy and non‐health science education studies, but not in nursing education and other disciplines in health science education (see Table 3 for all subgroup analyses results). The subgroup analysis by study design showed that the advantage of the FC condition was observed in cohort and quasi‐experimental studies, but not in RCTs. Of the four RCTs, only one conference study with small participant numbers from Thailand58 suggested an advantage for the FC approach.
Table 3

Subgroup analyses for post‐learning examination score outcomes stratified by various contexts

Subgroup and domainTest for heterogeneityTest for effect
I 2, % Q‐statisticp‐valuePooled effect size (95% CI)p‐value
Category of discipline (n = 41)
Medical (n = 14)80.566.59<0.0010.527 (0.311, 0.744)<0.001
Pharmacy (n = 7)92.580.39<0.0010.527 (0.122, 0.932)0.011
Nursing (n = 2)79.04.760.0290.196 (−0.333, 0.726)0.468
Other health science (n = 5)89.437.74<0.0010.200 (−0.343, 0.744)0.470
Non‐health science (n = 13)94.4216.19<0.0010.490 (0.143, 0.837)0.006
Study design (n = 41)
Cohort (n = 20)93.3284.74<0.0010.435 (0.173, 0.696)0.001
Quasi‐experimental (n = 17)87.2125.27<0.0010.548 (0.328, 0.768)<0.001
RCT (n = 4)54.66.610.0850.219 (−0.108, 0.545)0.190
Publication year (n = 41)
2012 (n = 1)0.858 (0.517,1.198)
2013 (n = 5)80.720.69<0.0010.168 (−0.205, 0.541)0.376
2014 (n = 7)89.556.99<0.0010.245 (−0.115, 0.606)0.182
2015 (n = 18)92.5226.45<0.0010.426 (0.167, 0.685)0.001
2016 (n = 10)81.849.58<0.0010.801 (0.526, 1.075)<0.001
Publication type (n = 41)
Journal (n = 32)92.2388.90<0.0010.458 (0.263, 0.653)<0.001
Thesis (n = 1)−0.018 (−0.379,0.344)
Conference (n = 8)72.025.000.0010.543 (0.306, 0.780)<0.001
Country (n = 41)
USA (n = 34)90.5348.56<0.0010.476 (0.304, 0.649)<0.001
Non‐USA (n = 7)91.973.65<0.0010.435 (−0.064, 0.935)0.088
Format of test (n = 41)
MCQ (n = 20)87.4150.97<0.0010.394 (0.186, 0.601)<0.001
Written essay (n = 3)93.731.93<0.0011.186 (0.451, 1.921)0.002
Combination (n = 5)51.68.260.0820.378 (0.174, 0.582)<0.001
Unclear (n = 13)93.9195.40<0.0010.451 (0.087, 0.815)0.015
Setting of medical education studies (n = 14)
School (n = 7)66.718.020.0060.578 (0.342,0.814)<0.001
Hospital (n = 7)85.541.42<0.0010.489 (0.123,0.856)0.009

CI = confidence interval; MCQ = multiple‐choice question; RCT = randomised controlled trial.

Subgroup analyses for post‐learning examination score outcomes stratified by various contexts CI = confidence interval; MCQ = multiple‐choice question; RCT = randomised controlled trial. For publication year, no significant differences were found in the outcome scores for 2013 and 2014, but higher examination scores for the FC condition were found in 2015 and 2016. Cumulative analysis by year (Fig. S7) revealed a progressive improvement in outcome scores for the FC condition over time. In terms of publication forum, both studies published in journals and conference studies showed a significantly better effect for the FC method. An advantage of the FC over the LB condition was present in studies originating from the USA, but not in studies from other countries. The advantage of the FC approach was also observed for all types of examination scores. Finally, we analysed the 14 medical education studies by setting. The FC condition had higher outcome scores than the LB condition in both hospital and school settings.

Meta‐regression for the examination score outcome measure

Meta‐regression by year (Fig. S8) showed a significant correlation (p = 0.023) between publication year and effect size (by excluding the year 2012, which had only one single study), but no significant relationship between sample size and outcome (p = 0.856).

Publication bias

A visual analysis via funnel plot (Fig. S9) revealed no obvious evidence of publication bias. Begg's test (p = 0.456) and Egger's regression test (p = 0.236) further confirmed no significant publication bias.

Discussion

We undertook a meta‐analysis of 46 studies comparing the FC and LB approaches to learning identified across a range of health science and non‐health science educational fields. To our knowledge, this is the most comprehensive meta‐analytical study to date on this topic. Our overall findings build on those from several recent systematic (narrative) reviews7, 8, 20, 33 and meta‐analyses40, 41 investigating the effectiveness of the FC in health care educational settings. Thus, the systematic narrative reviews reporting mixed findings from individual studies across medical and nursing education typically identified a relatively small number of studies and consistently called for more research and a wider reach in terms of examining the relative efficacy of the FC method across different educational groups.7, 8, 20, 33 Our meta‐analytical approach was adopted to avoid the subjectivity of the narrative approach to systematic reviews, and to pool the data from a large number of studies to ascertain the relative effectiveness of the FC method across a number of domains.59 Our work also builds on and extends the findings from the two meta‐analyses previously discussed.40, 41 Both of these studies found a significant effect for increased learning in Chinese nursing students using the FC method in terms of theoretical knowledge,40, 41 skills40, 41 and self‐directed learning.41 We included studies across all disciplines and countries, as long as the studies were available in the electronic databases we searched, verifying the findings from the Chinese studies for examination scores and course grades (i.e. related to theoretical knowledge) across a range of educational settings. However, for the OSCE outcome (related to skills) our analysis found no advantage for the FC condition. Despite the assertion that meta‐analysis can be undertaken using as few as two studies,57 this latter finding may be a result of the small number of studies reporting OSCE outcomes in our database and the fact that their results were greatly divergent. Our study also revealed inconsistencies in FC efficacy between different study designs: cohort or quasi‐experimental studies, but not RCTs, showed a significant advantage of the FC over the LB condition. This lack of efficacy for the FC condition found in our study directly contradicts the findings from the two earlier meta‐analyses,7, 8 which demonstrated a significant advantage of the FC approach in RCTs. This discrepancy may reflect, once again, the small number of RCTs in our dataset. The possibility of bias in the results of non‐RCT‐based studies should also be considered. Further differences were found in our results according to the particular student group under study. Thus, medical, pharmacy and non‐health science education student groups appeared to benefit from their FC experiences, but this was not the case for nursing and other health science students. Thus, our study has an advantage over the systematic reviews in medical education undertaken by Chen et al.7 and Hughes and Lyons,8 who concluded that results for the FC condition were mixed across studies, by pooling the data and gaining additional statistical power to ascertain the relative advantage of the FC approach within medical education settings. Indeed, the high statistical heterogeneity found in our meta‐analysis actually reflects the relative inconsistency between studies, as reported by Chen et al.7 and Hughes and Lyons.8 In terms of nurse education, however, our study fails to replicate the findings from the two meta‐analyses of the FC method using studies from the domain of nursing education in China. We are unsure exactly why this might be, although we suspect that it could be partly to do with the relative newness of the FC approach in those fields outside China,20, 33 alongside the paucity of studies in our analyses. Indeed, only three studies, distributed in different effect estimates, were incorporated in the nursing analysis. Furthermore, when considering the relative maturity of the FC methodology in each respective discipline, it is worth noting that our cumulative analyses suggest a tendency of progressive improvement in the outcome of the FC condition over time. Thus, progressive understanding of the FC method, the utilisation of newer technology, the developing maturity of teaching skills associated with the FC method and accumulated experience in the FC teaching format may all contribute to the improvement of outcomes.7 The results of our meta‐analyses have several limitations. For example, the studies had a high degree of statistical heterogeneity. It is likely that the statistical heterogeneity is a result of the large degree of methodological diversity. In this review methodological diversity exists and takes many forms, including, as discussed above, the implementation of the FC/LB conditions and the research design used to assess the differences. As such, although all studies addressed the implementation of an FC style of education, they varied in both the format and content of classroom flipping. For example, the experiences and expertise of teachers who designed and led the FC activity, and the types and qualities of pre‐class learning material were likely to differ. The content or cognitive level of knowledge evaluated (i.e. outcomes) also varied. The FC condition is designed to promote higher cognitive levels (e.g. application). As such, lower levels of assessment (e.g. recall) may not necessarily be considered as an appropriate outcome for the FC approach. The academic outcomes amongst our included studies mainly comprised three types: examination scores; OSCEs, and course grades. Most of these reflect a Kirkpatrick level 2 (learning) outcome.43 Additionally, the OSCE may also reflect a Kirkpatrick level 3 (behaviour) outcome and the MCQ may also include questions of a higher cognitive level than mere memorisation. As the formula for assessing course grades differed between studies (as we identified in Table S2), the cognitive levels evaluated also differed. Such diversity in study designs, outcomes and target populations contributes towards the heterogeneity in effect sizes. Given that heterogeneity was high, we must be cautious in drawing sweeping conclusions around the efficacy of the FC method. The issue of publication bias also needs to be addressed. We incorporated conference studies and a thesis in this review to avoid missing important information and publication bias. However, the methodological qualities of conference studies could not be assessed because of the limited information available. This gives these conference studies a high risk of bias. Despite this, the information obtained can be used as a reference. Given that the subgroup analysis showed a similar effect and direction of outcome between both conference and journal studies, we consider this to be positive support for our findings. Another limitation of our meta‐analysis is that the literature search was undertaken in June 2016 and because of the lengthy nature of undertaking such a detailed process (i.e. whittling down an initial 1706 identified contenders to the 46 included in this study, closely reading hundreds of papers written primarily in a non‐native language, data extraction, analyses and interpretation), it has taken us 18 months to deliver our results. As the number of publications on the effectiveness of the FC approach is rapidly growing, several new articles published after June 2016 were not included in our data pool. A new meta‐analysis 1 or 2 years later might yield further informative and significant findings. However, despite these limitations, our study has strengths. We have undertaken a meta‐analysis comprising data from 46 studies with a total of 9026 participants from different health care and non‐ health care educational disciplines across the world. This is not only the largest study of its kind that we know of, but, because it is a meta‐analysis, it provides a higher level of evidence60, 61 than do the descriptive review methods presently dominating the FC literature.7, 27, 33 Further, our approach has enabled us to consider the patterns within and across a variety of factors, including study population, year of study and research method, culminating in a greater understanding of the nuances around the FC method more generally than if we had restricted our data pool to one participant group alone. Our study leads us to make future research suggestions, focusing mainly on improving study methods, research design and reporting. Thus, a standard format of classroom flipping, with well‐trained and motivated facilitators, comprises the basic requirement for a comparison between FC and LB methods for future research studies. Further, given that the purpose of the FC method is to improve higher cognitive learning and promote behavioural change, greater attention should be given to ensuring that future research study designs are sensitive enough to measure the higher cognitive outcomes expected to delineate the efficacy of the FC. Indeed, the appraisal of a new pedagogy suggests that conceptual knowledge should be evaluated using novel tools, such as clinical reasoning measures.62 Furthermore, this paradigm shift in teaching methodology may require the development of a longitudinal assessment technique that measures clinical reasoning, high‐level cognitive performances and behaviour change to better realise the impact of this pedagogical intervention.2 Future research also needs to address the reporting of research: the poor reporting quality of educational studies is an important issue influencing the process of risk bias assessment. In many of the studies included in our meta‐analysis, a lot of necessary information was absent. For example, details of the number of eligible and enrolled participants, their demographic features, dropout rate, and the validity and reliability of assessment instruments were often missing. Age and gender are important confounders in clinical studies, and this is also the case in education research. However, many publications in our review failed to mention such demographics. Similar to our findings, Horsley et al.63 concluded that: ‘… reports of randomised studies in health professions education frequently omit elements recommended by the CONSORT statement. Most reports were assessed as having a high or unclear risk of bias.’ Thus, a structured format for reporting research in medical education studies is warranted. Without such rigour in the reporting of research studies, it will be impossible for us to implement a truly evidence‐based platform on which to make future decisions for curricula development.

Contributors

K‐SC and LVM contributed towards the design conceptualisation. All authors participated in the data acquisition, analysis and interpretation. K‐SC and Y‐HL performed the statistical analysis. K‐SC and LVM developed the first draft of the manuscript and led the critical revision of the paper. All authors approved the final manuscript.

Funding

none.

Conflicts of interest

none.

Ethical approval

not applicable. Figure S1. Meta‐analyses of the 32 health science studies by outcome measure. Figure S2. Meta‐analyses of the 14 non‐health science studies by outcome measure. Figure S3. Meta‐analyses of the medical education studies by outcome measure. Figure S4. Meta‐analyses of the pharmacy education studies by outcome measure. Figure S5. Meta‐analyses of the nursing education studies by outcome measure. Figure S6. Meta‐analyses of studies from other health science disciplines by outcome measure. Figure S7. Cumulative analysis of examination score outcome by publication year. Figure S8. Meta‐regression between examination scores and publication year. Figure S9. Funnel plot illustrating publication bias for the examination score outcome. Click here for additional data file. Table S1. Summary of the included studies. Table S2. Methodological qualities of journal studies by the Effective Public Health Practice Project (EPHPP) Assessment Tool. Click here for additional data file.
  42 in total

1.  How meta-analysis increases statistical power.

Authors:  Lawrence D Cohn; Betsy J Becker
Journal:  Psychol Methods       Date:  2003-09

Review 2.  A systematic review identifies a lack of standardization in methods for handling missing variance data.

Authors:  Natasha Wiebe; Ben Vandermeer; Robert W Platt; Terry P Klassen; David Moher; Nicholas J Barrowman
Journal:  J Clin Epidemiol       Date:  2006-04       Impact factor: 6.437

3.  The use of mean range as an estimator of variance in statistical tests.

Authors:  P B PATNAIK
Journal:  Biometrika       Date:  1950-06       Impact factor: 2.445

4.  Using the flipped classroom in graduate nursing education.

Authors:  Catharine M Critz; Diane Knight
Journal:  Nurse Educ       Date:  2013 Sep-Oct       Impact factor: 2.082

5.  The flipped classroom: a course redesign to foster learning and engagement in a health professions school.

Authors:  Jacqueline E McLaughlin; Mary T Roth; Dylan M Glatt; Nastaran Gharkholonarehe; Christopher A Davidson; LaToya M Griffin; Denise A Esserman; Russell J Mumper
Journal:  Acad Med       Date:  2014-02       Impact factor: 6.893

6.  Active learning and student-centered pedagogy improve student attitudes and performance in introductory biology.

Authors:  Peter Armbruster; Maya Patel; Erika Johnson; Martha Weiss
Journal:  CBE Life Sci Educ       Date:  2009       Impact factor: 3.325

7.  Rating the quality of evidence for clinical practice guidelines.

Authors:  D C Hadorn; D Baker; J S Hodges; N Hicks
Journal:  J Clin Epidemiol       Date:  1996-07       Impact factor: 6.437

8.  Flipping the Quality Improvement Classroom in Residency Education.

Authors:  Sara L Bonnes; John T Ratelle; Andrew J Halvorsen; Kimberly J Carter; Luke T Hafdahl; Amy T Wang; Jayawant N Mandrekar; Amy S Oxentenko; Thomas J Beckman; Christopher M Wittich
Journal:  Acad Med       Date:  2017-01       Impact factor: 6.893

9.  A Synthesis of Reading Interventions and Effects on Reading Comprehension Outcomes for Older Struggling Readers.

Authors:  Meaghan S Edmonds; Sharon Vaughn; Jade Wexler; Colleen Reutebuch; Amory Cable; Kathryn Klingler Tackett; Jennifer Wick Schnakenberg
Journal:  Rev Educ Res       Date:  2009-03-01

10.  Does the Concept of the "Flipped Classroom" Extend to the Emergency Medicine Clinical Clerkship?

Authors:  Corey Heitz; Melanie Prusakowski; George Willis; Christopher Franck
Journal:  West J Emerg Med       Date:  2015-10-22
View more
  28 in total

1.  Simulation-Based Mastery Learning Improves Patient and Caregiver Ventricular Assist Device Self-Care Skills: A Randomized Pilot Trial.

Authors:  Jeffrey H Barsuk; Jane E Wilcox; Elaine R Cohen; Rebecca S Harap; Kerry B Shanklin; Kathleen L Grady; Jane S Kim; Gretchen P Nonog; Lauren E Schulze; Alison M Jirak; Diane B Wayne; Kenzie A Cameron
Journal:  Circ Cardiovasc Qual Outcomes       Date:  2019-10-11

2.  Effective and Time-Efficient Implementation of a Flipped-Classroom in Preclinical Medical Education.

Authors:  Aaron M Marshall; Zachary E Conroy
Journal:  Med Sci Educ       Date:  2022-06-18

3.  Design and Validation of a Virtual Physical Education and Sport Science-Related Course: A Learner's Engagement Approach.

Authors:  Vijayamurugan Eswaramoorthi; Garry Kuan; Mohamad Razali Abdullah; Anwar P P Abdul Majeed; Pathmanathan K Suppiah; Rabiu Muazu Musa
Journal:  Int J Environ Res Public Health       Date:  2022-06-22       Impact factor: 4.614

4.  Evaluation of the effect of using online database to enhance students' learning of occupational therapy theories in occupational therapy education.

Authors:  Chi-Wen Chien; Chloe Mo; Chung-Ying Lin; Wing Hei Chan; Cheuk Wing Chan; Wai Hin Tang; Chun Pong Siu; Hang Shuen Yip
Journal:  Hong Kong J Occup Ther       Date:  2022-06-07       Impact factor: 1.476

5.  Development, Implementation, and Evaluation of an Open Access, Level-Specific, Core Content Curriculum for Emergency Medicine Residents.

Authors:  Kristen Grabow Moore; Andrew Ketterer; Natasha Wheaton; Paul Logan Weygandt; Holly A Caretta-Weyer; Jeremy Berberian; Jaime Jordan
Journal:  J Grad Med Educ       Date:  2021-10-15

6.  Effectiveness and Dissemination of the Interprofessional Pediatric Pain PRN Curriculum.

Authors:  Renee C B Manworren; Megan Basco
Journal:  J Contin Educ Health Prof       Date:  2021-12-01       Impact factor: 2.190

7.  Flipping the Classroom in Graduate Medical Education: A Systematic Review.

Authors:  Andrew M King; Michael Gottlieb; Jennifer Mitzman; Tina Dulani; Stephanie J Schulte; David P Way
Journal:  J Grad Med Educ       Date:  2019-02

8.  Flipped learning: Turning medical education upside down.

Authors:  Tahseen A Chowdhury; Halima Khan; Maralyn R Druce; William M Drake; Ravindra Rajakariar; Raj Thuraisingham; Hamish Dobbie; Laila Parvanta; Francis Chinegwundoh; Ahmad Almushatat; Anthony Warrens; Elspeth M Alstead
Journal:  Future Healthc J       Date:  2019-10

9.  Enhancing Qualities of Consciousness during Online Learning via Multisensory Interactions.

Authors:  Valentina Cesari; Benedetta Galgani; Angelo Gemignani; Danilo Menicucci
Journal:  Behav Sci (Basel)       Date:  2021-04-21

Review 10.  Mindfulness and Other Simple Neuroscience-Based Proposals to Promote the Learning Performance and Mental Health of Students during the COVID-19 Pandemic.

Authors:  Gonzalo R Tortella; Amedea B Seabra; Jorge Padrão; Rodrigo Díaz-San Juan
Journal:  Brain Sci       Date:  2021-04-27
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.