Literature DB >> 35125573

Testing the predictive accuracy of COVID-19 forecasts.

Laura Coroneo1, Fabrizio Iacone2,1, Alessia Paccagnini3,4, Paulo Santos Monteiro1.   

Abstract

We test the predictive accuracy of forecasts of the number of COVID-19 fatalities produced by several forecasting teams and collected by the United States Centers for Disease Control and Prevention for the epidemic in the United States. We find three main results. First, at the short horizon (1-week ahead) no forecasting team outperforms a simple time-series benchmark. Second, at longer horizons (3- and 4-week ahead) forecasters are more successful and sometimes outperform the benchmark. Third, one of the best performing forecasts is the Ensemble forecast, that combines all available predictions using uniform weights. In view of these results, collecting a wide range of forecasts and combining them in an ensemble forecast may be a superior approach for health authorities, rather than relying on a small number of forecasts.
© 2022 Published by Elsevier B.V. on behalf of International Institute of Forecasters.

Entities:  

Keywords:  Epidemic; Forecast evaluation; Forecasting tests

Year:  2022        PMID: 35125573      PMCID: PMC8801780          DOI: 10.1016/j.ijforecast.2022.01.005

Source DB:  PubMed          Journal:  Int J Forecast        ISSN: 0169-2070


Introduction

Forecasting the evolution of an epidemic is of utmost importance for policymakers and healthcare providers. Timely and reliable forecasts are necessary to help health authorities and the community at large cope with a surge of infections, and to inform public health interventions, for example, to enforce (or ease) a lockdown at the local or national level. Accordingly, in recent months a rapidly growing number of research teams have developed forecasts for the evolution of the current COVID-19 pandemic caused by the new coronavirus, SARS-CoV-2. In the United States, the Centers for Disease Control and Prevention (CDC) collects weekly forecasts of the evolution of the COVID-19 pandemic produced by different institutions and research teams. These forecasts are aimed at informing public health decision-making by projecting the probable impact of the COVID-19 pandemic at horizons up to 4 weeks. The forecasting teams that submit their forecasts to the CDC include data scientists, epidemiologists, and statisticians. They use different models and methods (e.g. SEIR, Bayesian, and deep learning models), combining a variety of data sources and assumptions about the impact of non-pharmaceutical interventions on the spread of the epidemic (such as social distancing and the use of face coverings). This wealth of forecasts can be extremely valuable for decision-makers, but it also poses a problem: how to act when confronted with heterogeneous forecasts and, in particular, how to select the most reliable projections. Decision-makers are thus faced with the task of comparing the predictive accuracy of different forecasts. Indeed, selecting models and comparing their predictive accuracy are different tasks, and in this paper we focus on the latter. As the Diebold and Mariano (DM) test of equal predictive accuracy (see Diebold and Mariano, 1995, Giacomini and White, 2006) adopts a model-free perspective to compare competing forecasts, imposing assumptions only on the forecast error loss differential, we use it to compare competing forecasts for the number of COVID-19 fatalities collected by the CDC. The application of the DM test is particularly challenging when only a few out-of-sample observations are available, as the standard test is unreliable, especially for multi-step forecasts (Clark & McCracken, 2013). To overcome this small-sample problem, we apply fixed-smoothing asymptotics, as recently proposed for this test by Coroneo and Iacone (2020). With fixed-smoothing asymptotics, the limit distribution of the DM statistic is derived under alternative assumptions. In particular, when the long-run variance in the test statistic is estimated as the weighted autocovariance estimate, the asymptotic distribution is derived assuming that the bandwidth-to-sample ratio (denoted as ) is constant, as recommended by Kiefer and Vogelsang (2005). With this alternative asymptotics, usually known as fixed-, the test of equal predictive accuracy has a nonstandard limit distribution that depends on and on the kernel used to estimate the long-run variance. The second alternative asymptotics that Coroneo and Iacone (2020) consider is the fixed- approach, as in Sun (2013) and Hualde and Iacone (2017). In this case, the estimate of the long-run variance is based on a weighted periodogram estimate, the asymptotic distribution is derived assuming that the truncation parameter is constant, and the test of equal predictive accuracy has a distribution with degrees of freedom that depend on the truncation parameter . Both approaches have been shown to deliver correctly sized predictive accuracy tests, even when only a small number of out-of-sample observations are available (see Coroneo and Iacone, 2020, Harvey et al., 2017). We evaluate forecasts for the cumulative number of COVID-19 fatalities produced at the national level for the United States by the eight forecasting teams that submitted their forecasts to the CDC without interruptions during the period from June 20, 2020 to March 20, 2021. Although the evaluation period includes only 40 observations, we document an increase in the volatility of the forecasting errors around the second half of the sample. Accordingly, we perform our forecast evaluation separately on two sub-samples of equal size: the first evaluation sub-sample (from June 20, 2020 to October 31, 2020) and the second evaluation sub-sample (from November 7, 2020 to March 20, 2021). This implies that for each evaluation sub-sample we can base our inference only on 20 observations, making the use of fixed-smoothing asymptotics crucial for obtaining reliable results. We compare the predictive accuracy of the forecasts of each team relative to the forecasts of a simple benchmark model, obtained by fitting a second-order polynomial using a rolling window of the last five available observations. We also consider two ensemble forecasts that combine the forecasts from several models using equal weights: one published by the CDC, and another one (the core ensemble) computed by us that combines only the forecasts included in our evaluation exercise. A feature that makes forecast evaluation important in its own right, especially when dealing with predicting the spread of COVID-19, is that the cost of under-predicting the spread of the disease can be greater than the cost of over-predicting it. In the midst of a public health crisis, the precautionary principle implies that erring on the side of caution is less costly than predicting the tapering off of the disease too soon. Scale effects may also be important in the evaluation of forecasts of an epidemic outbreak, since the same forecast error may be considered differently when the realized level of fatalities is small, and when there are a large number of fatalities. These effects may be taken into account in the forecast evaluation exercise by a judicious choice of the loss function. Therefore, we evaluate the predictive accuracy of each forecasting team using several loss functions, including the widely used quadratic and absolute value loss, the absolute percentage loss (that takes into account the scale of the number of fatalities), and a linear exponential loss function (that penalizes under-prediction more than over-prediction). Our findings can be summarized as follows. First, the simple polynomial benchmark outperforms the forecasters at the short horizon (1 week ahead), often significantly so. Second, at longer horizons (3 to 4 weeks ahead), the forecasters become more competitive, and some statistically outperform the simple benchmark, especially in the first evaluation sub-sample. This suggests that forecasters can successfully help inform forward-looking policy decisions. Third, the ensemble forecasts are among the best performing forecasts. This is particularly true in the first evaluation sub-sample, but even in the second sub-sample the ensemble forecast combinations outperform the benchmark, although in this sub-sample the DM test statistics are not statistically significant. The reliability of ensemble forecasts underlines the virtues of model averaging when uncertainty prevails, and supports the view in Manski (2020) that data and modeling uncertainties limit our ability to predict the impact of alternative policies using a tight set of models. Overall, our findings hold for all the loss functions considered and caution health authorities not to rely on a single forecasting team (or a small set) to predict the evolution of the pandemic. A better strategy appears to be to collect as many forecasts as possible and to use an ensemble forecast. The remainder of the paper is organized as follows. Section 2 lays out the methodology to implement the test of equal predictive accuracy. Section 3 describes the data and the models. Results are documented and discussed in Section 4, and Section 5 concludes. Finally, in the Online Appendix, we perform a Monte Carlo simulation exercise to study the size and power properties of the two tests of equal predictive ability with fixed-smoothing asymptotics for the sample sizes considered in our empirical study, and consider several additional experiments, including some alternative benchmark forecasts.

Forecast evaluation

We consider the time series of cumulative weekly deaths , with the sample size for which forecasts are available. We want to compare two -week-ahead forecasts, and , where for denotes the estimates obtained with a rolling window of size used to construct forecast , if known. The forecast error for forecast is and the associated loss is . For example, a quadratic loss would be and an absolute value loss would be . The null hypothesis of equal predictive ability of the two forecasts is Denote the time- loss differential between the two forecasts as and the sample mean of the loss differential as where . When a large sample is available, standard asymptotic theory may provide valid guidance for the statistical evaluation of ; see Diebold and Mariano (1995) and Giacomini and White (2006). However, the same inference may be severely biased when the sample has only a moderate size, as it is indeed the case when comparing the forecast accuracy of predictions of the number of fatalities of COVID-19. In this case, fixed- and fixed- asymptotics can be used to overcome the small-sample size bias; see Coroneo and Iacone (2020), Choi and Kiefer (2010), and Harvey et al. (2017). As for the fixed- asymptotics, following Kiefer and Vogelsang (2005), under the null in (1), where denotes the weighted autocovariance estimate of the long-run variance of using the Bartlett kernel and truncation lag . Kiefer and Vogelsang (2005) characterize the limit distribution and provide formulas to compute quantiles. For example, for the Bartlett kernel with , these can be obtained using the formula where When testing assumptions about the sample mean, Kiefer and Vogelsang (2005) show in Monte Carlo simulations that the fixed- asymptotics yields a remarkable improvement in size. However, while the empirical size improves (it gets closer to the theoretical size) as is closer to 1, the power of the test worsens, implying that there is a size–power tradeoff. For fixed- asymptotics, following Hualde and Iacone (2017), under the null in (1), we have where is the weighted periodogram estimate of the long-run variance of using the Daniell kernel and truncation . Similar results, with a slightly different standardization and therefore a slightly different limit, are in Sun (2013). Monte Carlo simulations in Hualde and Iacone (2017) and Lazarus, Lewis, Stock, and Watson (2018) show that fixed- asymptotics has the same size–power tradeoff documented for fixed- asymptotics: the smaller the value for , the better the empirical size, but also the weaker the power. Coroneo and Iacone (2020) analyze the size and power properties of the tests of equal predictive accuracy in (2), (3) in an environment with asymptotically non- vanishing estimation uncertainty, as in Giacomini and White (2006). The results indicate that the tests in (2), (3) deliver correctly sized predictive accuracy tests for correlated loss differentials even in small samples, and that the power of these tests mimics the size-adjusted power. Considering size control and power loss in a Monte Carlo study, they recommend the bandwidth for the weighted autocovariance estimate of the long-run variance using the Bartlett kernel (where denotes the integer part of a number) and for the weighted periodogram estimate of the long-run variance using the Daniell kernel. In Appendix A, we perform a Monte Carlo simulation exercise to investigate the empirical size and power of the two tests for sample sizes that match the ones in our empirical study. We simulate forecast errors as in Clark, 1999, Coroneo and Iacone, 2020, Diebold and Mariano, 1995. Our findings indicate that both tests are, in general, correctly sized, even when only 20 observations are available and in presence of autocorrelation of the loss differential, although the test with WCE and fixed- asymptotics can be slightly oversized in short samples and in presence of strong autocorrelation. On the other hand, the test with WPE and fixed- asymptotics trails slightly behind the test with WCE in terms of power.

Forecasting teams and benchmark

Data and forecasting teams

In our empirical investigation, we use forecasts for the cumulative number of deaths collected by the Centers for Disease Control and Prevention (CDC). The CDC is a federal agency in charge of protecting public health through the control and prevention of diseases. It is also the official source of statistics on the COVID-19 pandemic evolution in the US. In particular, in collaboration with independent teams of forecasters, the CDC has set up a repository of weekly forecasts for the numbers of deaths, hospitalizations, and cases. These forecasts are developed independently by each team and shared publicly.1 We focus on forecasts of the number of deaths for three main reasons. First, the number of fatalities is more accurate than the number of cases and hospitalizations, since the latter ignores asymptomatic cases and other diseases that are undetected. Second, the number of deaths is reported with less spatial and temporal biases. Third, when faced with a pandemic, the number of fatalities is arguably the primary concern of the health authorities and of the public. Our sample includes projections for national COVID-19 cumulative deaths made for the period between June 20, 2020 and March 20, 2021 by eight forecasting teams. The deadline for the teams to submit their weekly forecasts is on the Monday of each week, and they are usually published online on Wednesdays. Weekly cumulative data are the cumulative data up to and including Saturday. This means that, for example, the forecasts submitted by June 22 had as targets the cumulative number of deaths as of June 27 (1 week ahead), July 2 (2 weeks ahead), July 7 (3 weeks ahead), and July 12 (4 weeks ahead). Realized values are also taken from the CDC website. Notice that when COVID-19 is reported as a cause of mortality on the death certificate, it is coded and counted as a fatality due to COVID-19. The eight forecasting teams selected are those that submitted their predictions with no interruptions for all the weeks in our sample. We list the selected teams in Table 1 and report the main features of the selected forecasts. They vary widely with regards to their modeling choice, information input (for example, how the information on infected people is used), and in their assumptions about the evolution and impact of non-pharmaceutical interventions (for example regarding social distancing).2
Table 1

Forecasting teams, methods, and assumptions.

CodeTeamModelMethodChange
COCOVID Analytics – MIT SloanDELPHI ModelDeep learning modelno
UMUniversity of Massachusetts, AmherstUMass – MBMechanistic Bayesian compartment modelno
UAUniversity of ArizonaUA – EpiCovDAModified SEIR modelyes
GTGeorgia Institute of Technology, Deep Outbreak ProjectGT – Deep COVIDDeep learning modelno
MONortheastern University, Laboratory for the Modeling of Biological and Socio-technical SystemsMOBS – GLEAM COVIDMetapopulation, age-structured SEIR modelno
PSPredictive Science, Inc.PS – DRAFTSEIR modelyes
LALos Alamos National LaboratoryLANL – Growth RateStatistical dynamical growth modelno
JHJohns Hopkins University, Infectious Disease Dynamics LabJHU – IDD – CovidSPMetapopulation SEIR modelyes

Notes: The column code describes the code given in the empirical analysis to each team. A forecasting team is included if it submitted its predictions for all the weeks in our sample. The table reports for each forecasting team the modeling methodology and whether the model considers a change in the assumptions about policy interventions. In the fourth column, “yes“ means that the modeling team makes assumptions about how levels of social distancing will change in the future, while “no” means that it is assumed that the existing measures will continue through the projected 4-week time period.

Forecasting teams, methods, and assumptions. Notes: The column code describes the code given in the empirical analysis to each team. A forecasting team is included if it submitted its predictions for all the weeks in our sample. The table reports for each forecasting team the modeling methodology and whether the model considers a change in the assumptions about policy interventions. In the fourth column, “yes“ means that the modeling team makes assumptions about how levels of social distancing will change in the future, while “no” means that it is assumed that the existing measures will continue through the projected 4-week time period.

Ensemble forecasts

In our forecast evaluation exercise, we also consider two ensemble forecasts: one published by the CDC (that combines the individual forecasts from several teams and that we label EN) and one computed by us (that combines the individual forecasts from the eight teams listed in Table 1 and that we call the core ensemble, CE).3 Combining forecasts is an effective procedure when there is uncertainty about the model and the data, as it is indeed the case here, where differences also include alternative assumptions on the responses of the public and of the health authorities. In this situation, combining forecasts is useful, as it helps to diversify risk and to pool information (see Bates & Granger, 1969). In particular, forecast combination is most advantageous when there is pervasive uncertainty, as the ranking of best performing forecasts may be very unstable and therefore forecast combination provides a robust alternative (see Stock and Watson, 1998, Timmermann, 2006). Optimal weights that give the best combination, in the sense of minimizing a given loss function, can actually be derived, but in many practical applications, estimated optimal weights schemes result in a combined forecast that does not improve simple averaging (see Claeskens et al., 2016, Clemen, 1989, Smith and Wallis, 2009). In epidemiology, forecast combination has proved its ability to improve on the performance of individual competing models. For example, Reich et al. (2019) found that ensemble forecasting for influenza performed better on average against the constituting models; similar results were also obtained by Chowell et al. (2020) in the Ebola Forecasting Challenge. Both of these works had access to a sufficiently long history of data, making a data-driven selection of the weights assigned to the contributing models possible. Interestingly, Reich et al. (2019) also considered the equal-weighting scheme in their exercise, and found that this naive ensemble performed quite well even against the one with data-driven weights, making it a reasonable choice for the current situation of a new epidemic, in which no previous outbreaks exist and no previous track record of past models is available. The ensemble forecast produced by the CDC is also naive, in the sense that it treats equally all the available forecasts. Specifically, it is obtained by combining the forecasts of all the teams that submitted to the CDC, as long as they submitted forecasts up to 4 weeks ahead and these forecasts were at least equal to the level observed on the day the forecast was submitted. The weekly composition of the pool of models contributing to the CDC ensemble forecast changes, and it includes, in general, a larger number of teams than the one we consider in our evaluation exercise.4 This loose criterion allows to include as many forecasts as possible, which may be desirable, but there is also the risk of including poorly performing teams. For this reason, we also consider a core ensemble constructed by us, which uses only the forecasts (equally weighted) by the eight teams that are included in our forecast evaluation exercise. The conjecture motivating this choice is that, as these are the most long-standing forecasting teams, they should also be the most experienced. This experience may give them an edge relative to other teams. In addition, by comparing the performance of the individual forecasts with the core ensemble forecast, we can reliably assess the value added by the combination of the forecasts, as the core ensemble uses only forecasts that are included in our exercise.

Benchmark forecasts

The benchmark against which we compare the forecasts collected by the CDC is a polynomial function. That is, benchmark forecasts are obtained as projections from the model: where is the cumulative number of fatalities, is the time trend, and is an unobserved error term. To accommodate the fact that the forecasted patterns may need changing even over a short span of time, we fit the quadratic polynomial model using least squares with a rolling window of the last five observations (using weekly data, this covers approximately a month). To ensure that the benchmark forecasts for the cumulative number of deaths are not decreasing, we compute the benchmark predictions as the maximum between the current value and the prediction from (4). This very simple statistical model has been chosen because any continuous and differentiable function can be approximated locally by a polynomial, and we take the second degree polynomial as a local approximation. In recent works, the choice of a polynomial benchmark has also been considered by Jiang, Zhao, and Shao (2020) and Li and Linton (2021), among others, although with some small differences. In Jiang et al. (2020), the intrinsic instability of the forecasted patterns is accommodated by fitting occasional breaks; Li and Linton (2021) fitted the model to the incidence of deaths, rather than to the cumulative deaths.

Preliminary analysis

In this section, we present some preliminary analysis of the forecasts submitted by the forecasting teams in Table 1, the ensemble (EN) forecast published by the CDC, the core ensemble (CE) constructed by combining all the forecasts of the teams in Table 1, and the forecast of the polynomial benchmark (PO), described above. Fig. 1 plots all the forecasts considered for the 1- to 4-week-ahead forecasting horizons, alongside the realized data. Comparing the graphs in Fig. 1 at different horizons, it is apparent that the heterogeneity in forecasts grows with the forecasting horizon and, concurrently, that the forecasts are less precise as the forecast horizon increases. This simple observation may make the case for forecast combination at longer horizons more compelling.
Fig. 1

Cumulative deaths in the US, observed vs. forecasted.

Note: Forecasts at forecasting horizons from 1 to 4 weeks, along with realized cumulative fatalities. Weekly observations from June 20, 2020 to March 20, 2021. The vertical line indicates November 3, 2020 and delimits the two sub-samples. The names of the forecasting teams are as in Table 1; EN denotes the ensemble published by the CDC, CE denotes the core ensemble constructed by combining all the forecasts of the teams in Table 1, and PO denotes the polynomial benchmark.

Fig. 2 plots the forecast errors for each forecasting team, the ensembles, and the benchmark (computed as the difference between the realization and the point forecast). The figure indicates that most forecasting teams seem to have systematically under-predicted the target, in particular in the second part of the sample. This is, of course, relevant for policymakers if the costs of over-prediction and under-prediction are different. Fig. 2 also shows that the size of the forecast errors is increasing with the forecasting horizon, suggesting that there is more uncertainty about the evolution of the epidemic in the long run (4 weeks ahead), compared to the short run (1 week ahead).
Fig. 2

Forecast errors.

Note: Forecast errors at forecasting horizons from 1 to 4 weeks. Weekly observations from June 20, 2020 to March 20, 2021. The vertical line indicates November 3, 2020 and delimits the two sub-samples. The names of the forecasting teams are as in Table 1; EN denotes the CDC ensemble forecast, CE denotes the core ensemble, and PO the polynomial benchmark. Forecast errors are defined as the realized value minus the forecast.

Cumulative deaths in the US, observed vs. forecasted. Note: Forecasts at forecasting horizons from 1 to 4 weeks, along with realized cumulative fatalities. Weekly observations from June 20, 2020 to March 20, 2021. The vertical line indicates November 3, 2020 and delimits the two sub-samples. The names of the forecasting teams are as in Table 1; EN denotes the ensemble published by the CDC, CE denotes the core ensemble constructed by combining all the forecasts of the teams in Table 1, and PO denotes the polynomial benchmark. Forecast errors. Note: Forecast errors at forecasting horizons from 1 to 4 weeks. Weekly observations from June 20, 2020 to March 20, 2021. The vertical line indicates November 3, 2020 and delimits the two sub-samples. The names of the forecasting teams are as in Table 1; EN denotes the CDC ensemble forecast, CE denotes the core ensemble, and PO the polynomial benchmark. Forecast errors are defined as the realized value minus the forecast. Incident deaths in the US over the evaluation period. Note: Weekly observations from June 20, 2020 to March 20, 2021. The vertical line indicates November 3, 2020 and delimits our two evaluation sub-samples. Table 2 presents some summary statistics for the forecast errors. The table reports for each forecasting horizon and forecasting scheme (team, ensemble, core ensemble, or polynomial) the sample mean, median, standard deviation, skewness, and the sample autocorrelations up to order 4 (AC(1), AC(2), AC(3), AC(4)). With the exception of the benchmark, the average of the forecast errors is positive for all forecasts, meaning that the forecasters tend to under-predict the fatalities.
Table 2

Summary statistics of forecast errors.

1 week aheadMeanMedianStdSkewAC(1)AC(2)AC(3)AC(4)
CO2475.851920.004429.680.220.190.060.150.43
UM3363.881705.503693.801.190.830.720.620.52
UA3663.352062.004310.210.870.730.650.550.51
GT2695.211334.093651.890.710.430.400.380.39
MO2520.661549.872965.750.900.700.560.520.54
PS3057.912454.003028.030.920.700.460.430.38
LA2860.782108.792883.871.310.470.220.230.40
JH5096.852333.364921.480.750.780.610.600.71
EN2754.321678.503246.071.110.710.540.520.55
CE3216.812015.613256.361.100.740.560.560.62
PO−97.6691.002210.91−0.740.22−0.44−0.160.35

2 weeks aheadMeanMedianStdSkewAC(1)AC(2)AC(3)AC(4)

CO3737.882200.005428.870.870.420.110.250.44
UM3830.681697.004541.001.520.700.430.520.59
UA4586.732705.506911.650.980.710.520.510.52
GT3330.982166.464945.810.250.310.200.360.35
MO3004.342365.753780.380.810.560.260.330.44
PS3759.053333.505381.690.600.560.150.140.19
LA4128.153400.924235.161.370.16−0.14−0.060.35
JH4986.192812.736491.230.490.680.440.470.61
EN3322.701808.504459.981.270.610.250.360.47
CE3920.502627.344299.011.280.660.330.470.60
PO−89.381053.405021.96−0.750.42−0.26−0.140.13

3 weeks aheadMeanMedianStdSkewAC(1)AC(2)AC(3)AC(4)

CO4971.903164.007114.770.880.600.270.250.36
UM4410.382082.005849.592.000.520.220.350.58
UA5644.753893.509930.941.200.700.480.430.51
GT3988.102844.716407.18−0.200.250.260.370.32
MO3496.542592.725166.810.470.480.270.050.20
PS4129.983890.759066.45−0.030.520.090.060.06
LA5992.515576.285790.070.720.05−0.20−0.180.28
JH4343.702386.358874.590.430.710.550.530.60
EN3951.132309.005562.791.320.600.280.210.35
CE4622.233005.155569.311.320.660.350.390.54
PO27.632232.548841.85−0.870.580.01−0.130.01

4 weeks aheadMeanMedianStdSkewAC(1)AC(2)AC(3)AC(4)

CO6594.934090.009744.800.700.730.360.260.26
UM4963.633037.507521.182.130.430.010.180.44
UA7021.954854.0012912.431.370.720.440.410.48
GT4875.403314.538662.06−0.140.360.180.370.30
MO3882.063566.277309.12−0.110.540.30−0.010.01
PS4058.405026.7513878.24−0.500.500.110.050.01
LA8273.187802.858444.130.060.16−0.06−0.080.37
JH2769.8216.3112755.550.540.750.630.600.63
EN4641.082610.507286.161.160.660.310.220.29
CE5304.922759.807359.801.110.700.370.370.45
PO−301.204249.0313860.09−1.030.640.180.020.07

Notes: The table reports summary sample statistics of forecast errors for the teams, the ensemble (EN), the core ensemble (CE), and the polynomial (PO) forecasts. The table the reports mean, median, standard deviation (std), skewness (skew), and autocorrelation coefficients up to order 4 (AC(1), AC(2), AC(3), AC(4)). Weekly observations from June 20, 2020 to March 20, 2021.

Summary statistics of forecast errors. Notes: The table reports summary sample statistics of forecast errors for the teams, the ensemble (EN), the core ensemble (CE), and the polynomial (PO) forecasts. The table the reports mean, median, standard deviation (std), skewness (skew), and autocorrelation coefficients up to order 4 (AC(1), AC(2), AC(3), AC(4)). Weekly observations from June 20, 2020 to March 20, 2021. At the 1-week horizon, the benchmark polynomial model appears to outperform all the teams, with a much smaller average error and smaller dispersion. However, its performance deteriorates at longer horizons, with the volatility of the forecast errors increasing substantially, and becoming larger than those of most forecasting teams. This is not surprising, as epidemiological models are designed to predict the evolution of a pandemic in the medium and the long run, and we observe here that even a very simple forecast does better when the horizon is very short. At longer horizons, however, epidemiological models should be expected to produce forecasts that are superior to simple statistical benchmarks. From Table 2 we can also observe that the forecast errors are autocorrelated, as documented in columns AC(1), AC(2), AC(3), and AC(4). This happens even for one-step-ahead forecasts, where the first-order sample autocorrelation may be as high as 0.83. This is interesting because, under the mean squared error loss, optimal -step ahead forecast errors should be at most MA(): so -step-ahead forecast errors should be Martingale differences, 2-step-ahead errors should be at most MA(1), and so on. Indeed, this is the very argument given in Diebold and Mariano (1995) to justify the choice of the rectangular kernel to estimate the long-run variance. However, this condition is clearly violated by all forecasts. One explanation of this higher-order autocorrelation in the forecast errors and the fact that the forecasting teams systematically under-predict the number of fatalities could be that the forecasting teams use alternative loss functions to produce their forecasts. Indeed, Patton and Timmermann (2007) show that, under asymmetric loss and nonlinear data generating processes, forecast errors can be biased and serially correlated of arbitrarily high order. Finally, Fig. 2 shows a break in the volatility of the forecast errors across the first and the second halves of the sample (as illustrated by the vertical line in each diagram of Fig. 1, Fig. 2). This is also shown in Table 3, Table 4, where we report summary statistics for the forecast errors in the two sub-samples. In particular, we note that the volatility of the forecast errors is considerably higher in the second sub-sample. Such decline in the quality of the forecasts in the most recent sub-sample may at first be puzzling: one would expect the forecasting teams to improve their performance as more information becomes progressively available. However, this structural break in the forecasting ability of all models could in part be related to the emergence of a new strain of the virus at the end of 2020, with specific mutations in the spike protein of SARS-CoV-2 resulting in increased transmissibility. Consistent with this explanation, research from the CDC reports that the B.1.1.7 virus strain (often referred to as the “Kent” variant) is estimated to have emerged in September 2020. This variant exhibited rapid growth in the US in early 2021, and was predicted to be the predominant variant by March 2021 (Galloway et al., 2021). This is illustrated in Fig. 3, showing the incident deaths across the overall evaluation period. Thus the increased forecasting errors may be driven by the sudden heightened incident deaths and associated increased slope of the cumulative deaths curve.
Table 3

Summary statistics of forecast errors – first evaluation sub-sample.

1 week aheadMeanMedianStdSkewAC(1)AC(2)AC(3)AC(4)
CO965.101060.501238.58−0.590.61−0.13−0.57−0.64
UM1059.70740.00799.460.990.33−0.09−0.02−0.12
UA1149.25718.502377.892.200.010.20−0.030.01
GT961.88340.842235.863.32−0.140.200.18−0.09
MO860.03804.30914.070.10−0.010.180.19−0.15
PS1461.631469.501388.680.120.12−0.170.120.20
LA1583.011261.771034.430.680.210.13−0.26−0.26
JH1314.881135.981028.010.210.600.16−0.13−0.35
EN878.69781.00702.74−0.080.220.100.15−0.23
CE1169.431072.08739.990.240.240.330.20−0.11
PO184.05240.70863.85−0.010.520.100.11−0.18

2 weeks aheadMeanMedianStdSkewAC(1)AC(2)AC(3)AC(4)

CO1248.001622.002072.00−0.760.600.05−0.24−0.39
UM1286.201245.50933.210.700.330.05−0.22−0.32
UA535.85−223.003737.170.230.230.320.01−0.13
GT1099.22436.942391.342.65−0.140.34−0.010.08
MO1334.241707.041769.66−1.580.060.18−0.16−0.09
PS1766.032284.752986.82−1.130.25−0.27−0.080.17
LA2873.012637.792117.380.180.410.11−0.20−0.45
JH−24.23−308.801874.640.380.680.32−0.18−0.40
EN988.15653.501227.020.410.370.14−0.23−0.33
CE1264.791296.371069.340.250.400.24−0.11−0.15
PO665.811050.202286.05−0.300.720.310.13−0.11

3 weeks aheadMeanMedianStdSkewAC(1)AC(2)AC(3)AC(4)

CO1733.152588.002487.69−1.110.49−0.04−0.10−0.31
UM1543.351401.001212.550.030.150.16−0.07−0.26
UA−307.65−1466.005456.380.260.360.400.01−0.20
GT1179.99869.562787.102.08−0.070.45−0.04−0.01
MO1761.962198.233470.14−1.610.170.19−0.22−0.25
PS1303.082203.005607.86−1.790.40−0.19−0.18−0.08
LA4844.444477.223177.370.130.510.15−0.25−0.50
JH−2536.57−2703.053367.77−0.010.800.490.16−0.08
EN1151.16648.451792.830.710.500.25−0.20−0.38
CE1190.22948.761733.000.160.480.16−0.15−0.38
PO1546.591622.034607.05−0.360.810.460.20−0.05

4 weeks aheadMeanMedianStdSkewAC(1)AC(2)AC(3)AC(4)

CO2495.203046.002746.48−0.530.46−0.14−0.11−0.24
UM1985.652532.002120.85−0.99−0.09−0.010.18−0.05
UA−862.95−3495.007232.630.730.470.430.14−0.22
GT1346.80293.713724.081.700.040.44−0.02−0.02
MO2238.443443.735868.38−1.700.290.22−0.25−0.33
PS200.981934.759607.37−1.970.48−0.08−0.23−0.22
LA6953.066431.774444.550.160.560.320.01−0.32
JH−6589.80−5408.155803.66−0.380.850.660.280.07
EN1309.30368.502490.151.140.520.37−0.01−0.34
CE970.92597.72736.020.360.560.25−0.13−0.56
PO2672.812938.397805.47−0.350.830.530.25−0.04

Notes: The table reports summary statistics of forecast errors for the teams, the ensemble (EN), the core ensemble (CE), and the polynomial (PO) forecasts. The table reports the mean, median, standard deviation (std), skewness (skew), and autocorrelation coefficients up to order 4 (AC(1), AC(2), AC(3), AC(4)). Weekly observations from June 20, 2020 to October 31, 2021.

Table 4

Summary statistics of forecast errors – second evaluation sub-sample.

1 week aheadMeanMedianStdSkewAC(1)AC(2)AC(3)AC(4)
CO3986.604375.005825.61−0.560.01−0.130.010.38
UM5668.055257.004023.670.220.710.500.280.04
UA5668.055257.004023.670.220.710.500.280.04
GT4428.554503.654006.28−0.380.380.200.160.14
MO4181.284291.583378.33−0.130.580.300.240.27
PS4654.204090.503395.220.150.650.270.170.05
LA4138.553610.503544.550.430.33−0.050.000.24
JH8878.819467.994306.70−0.210.43−0.01−0.050.21
EN4629.954108.503705.170.050.560.250.220.27
CE5264.194893.073520.450.100.560.180.180.28
PO−379.36−446.503019.97−0.340.19−0.51−0.220.36

2 weeks aheadMeanMedianStdSkewAC(1)AC(2)AC(3)AC(4)

CO6227.756081.006569.00−0.010.21−0.190.050.34
UM6375.154424.505275.000.570.540.090.260.31
UA8637.607281.007038.680.630.630.150.200.21
GT5562.745781.505831.40−0.840.15−0.120.220.15
MO4674.444429.204508.980.130.48−0.010.160.36
PS5752.085342.256493.530.080.520.04−0.010.00
LA5383.284634.005386.770.770.01−0.35−0.170.39
JH9996.6010425.315488.58−0.810.12−0.48−0.280.13
EN5657.255428.005277.330.310.44−0.120.110.26
CE6576.205324.994684.470.400.42−0.190.120.35
PO−844.561196.906733.56−0.330.37−0.37−0.250.11

3 weeks aheadMeanMedianStdSkewAC(1)AC(2)AC(3)AC(4)

CO8210.658926.508696.68−0.020.490.060.030.24
UM7277.405187.007173.651.060.36−0.070.100.43
UA11597.1511219.509903.231.180.560.040.030.25
GT6796.228185.657739.23−1.300.040.000.190.17
MO5231.134332.776034.960.330.500.15−0.050.22
PS6956.889382.5010975.21−0.380.490.03−0.02−0.05
LA7140.597402.157479.550.30−0.10−0.34−0.230.39
JH11223.9810597.457117.63−0.590.13−0.29−0.270.17
EN6751.106560.006618.050.410.45−0.01−0.120.15
CE8054.257246.775988.860.580.43−0.11−0.030.34
PO−1491.333231.8011592.63−0.440.52−0.12−0.26−0.04

4 weeks aheadMeanMedianStdSkewAC(1)AC(2)AC(3)AC(4)

CO10694.6514709.5012328.52−0.250.670.210.070.12
UM7941.604073.509640.821.210.34−0.200.000.34
UA14906.8512236.5012611.041.620.56−0.08−0.060.24
GT8404.019258.7810673.80−1.120.26−0.160.240.21
MO5525.684118.238338.640.140.630.26−0.010.11
PS7915.8312412.5016483.96−0.700.450.060.01−0.05
LA9593.3010939.2611087.67−0.260.09−0.15−0.120.45
JH12129.4411167.0210763.970.020.340.010.050.30
EN7972.857476.508910.650.230.570.06−0.040.13
CE9638.929422.568009.630.310.53−0.040.010.28
PO−3275.216527.3917741.72−0.580.580.05−0.090.03

Notes: The table reports summary statistics of forecast errors for the teams, the ensemble (EN), the core ensemble (CE), and the polynomial (PO) forecasts. The table reports the mean, median, standard deviation (std), skewness (skew), and autocorrelation coefficients up to order 4 (AC(1), AC(2), AC(3), AC(4)). Weekly observations from November 7, 2020 to March 20, 2021.

Fig. 3

Incident deaths in the US over the evaluation period.

Note: Weekly observations from June 20, 2020 to March 20, 2021. The vertical line indicates November 3, 2020 and delimits our two evaluation sub-samples.

Summary statistics of forecast errors – first evaluation sub-sample. Notes: The table reports summary statistics of forecast errors for the teams, the ensemble (EN), the core ensemble (CE), and the polynomial (PO) forecasts. The table reports the mean, median, standard deviation (std), skewness (skew), and autocorrelation coefficients up to order 4 (AC(1), AC(2), AC(3), AC(4)). Weekly observations from June 20, 2020 to October 31, 2021. Summary statistics of forecast errors – second evaluation sub-sample. Notes: The table reports summary statistics of forecast errors for the teams, the ensemble (EN), the core ensemble (CE), and the polynomial (PO) forecasts. The table reports the mean, median, standard deviation (std), skewness (skew), and autocorrelation coefficients up to order 4 (AC(1), AC(2), AC(3), AC(4)). Weekly observations from November 7, 2020 to March 20, 2021. At any rate, the volatility of the forecasting errors increases markedly starting from the beginning of November 2020. We thus perform our forecast evaluation separately on two equally sized sub-samples: the first evaluation sub-sample (from June 20, 2020 to October 31, 2020), and the second evaluation sub-sample (from November 7, 2020 to March 20, 2021). This means that for each evaluation sub-sample we base our inference on just 20 observations. With such small sample sizes, fixed-smoothing asymptotics is crucial to obtain correctly sized tests.

Forecast evaluation results

Our main results for the test of equal predictive ability of each forecasting team vis-à-vis the benchmark model (4) are reported in Table 5.
Table 5

Tests for equal predictive ability.

Note: The table reports test statistics for the test of equal predictive accuracy using the weighted covariance estimator (WCE) and the weighted periodogram estimator (WPE) of the long-run variance. The benchmark is a second-degree polynomial fitted on a rolling window of five observations. The forecast errors are evaluated using the absolute value loss function. A positive value of the test statistic indicates lower loss for the forecaster (i.e. better performance of the forecaster relative to the polynomial model). and respectively indicate two-sided significance at the 5% and 10% levels using fixed- asymptotics for WCE and fixed- asymptotics for WPE. and respectively indicate two-sided significance at the 5% and 10% levels using the bootstrap. Bootstrap critical values are constructed using the overlapping stationary block-bootstrap of Politis and Romano (1994), using an average block length of and a circular scheme, as described in Coroneo and Iacone (2020).

We conduct the analysis separately for the two evaluation sub-samples identified in Fig. 2. This yields 20 observations for each sub-sample, underlying the importance of alternative asymptotics in evaluating predictive ability.5 In the baseline analysis, we evaluate forecast errors relying on the absolute value loss function. We present the test statistics using both the weighted covariance estimator with the Bartlett kernel (WCE) and the weighted periodogram estimator with the Daniell kernel (WPE) of the long-run variance. A positive value for the test statistic indicates that the forecast in question is more accurate than the benchmark. We report two-sided significance at the 5% and 10% levels, using fixed-smoothing asymptotics (fixed- for WCE and fixed- for WPE) to establish the critical values. In particular, for and the bandwidth recommendations in Coroneo and Iacone (2020), the critical values are and with fixed- asymptotics, and and with fixed- asymptotics. For comparison, we also report significance based on bootstrap critical values, constructed using the overlapping stationary block-bootstrap of Politis and Romano (1994), using an average block length of and a circular scheme, as described in Coroneo and Iacone (2020). Tests for equal predictive ability. Note: The table reports test statistics for the test of equal predictive accuracy using the weighted covariance estimator (WCE) and the weighted periodogram estimator (WPE) of the long-run variance. The benchmark is a second-degree polynomial fitted on a rolling window of five observations. The forecast errors are evaluated using the absolute value loss function. A positive value of the test statistic indicates lower loss for the forecaster (i.e. better performance of the forecaster relative to the polynomial model). and respectively indicate two-sided significance at the 5% and 10% levels using fixed- asymptotics for WCE and fixed- asymptotics for WPE. and respectively indicate two-sided significance at the 5% and 10% levels using the bootstrap. Bootstrap critical values are constructed using the overlapping stationary block-bootstrap of Politis and Romano (1994), using an average block length of and a circular scheme, as described in Coroneo and Iacone (2020). MAE across sub-samples. Notes: The table reports the MAE of forecast errors for each team, the ensemble (EN), the core ensemble (CE), and the polynomial (PO) forecasts. The top panel shows the MAE level, and the bottom panel shows the MAE relative to the MAE of the benchmark model. The first evaluation sub-sample is from June 20, 2020 to October 31, 2020, and the second evaluation sub-sample from November 7, 2020 to March 20, 2021. We first consider the upper-panel of Table 5, which reports results for the first sub-sample from June 20, 2020 to October 31, 2020. The results indicate that no forecasting scheme predicts better than the benchmark at the 1-week forecasting horizon. In fact, we find that the benchmark often significantly outperforms the forecasting teams. On the other hand, at the 2-week horizon, the sign of some test statistics turns from negative to positive, reflecting a smaller relative loss by the forecasting teams. The relative performance of the forecasters improves further at longer horizons (3 and 4 weeks ahead), and we observe statistically significant relative gains in performance for some forecasting teams and the ensemble forecasts. The MAEs for the first sub-sample reported in Table 6 indicate that, at 3 and 4 weeks ahead, both ensemble forecasts (the EN ensemble provided by the CDC and the CE core ensemble constructed by combining all the forecasts of the teams in Table 1) performed better than the individual forecasting teams. This finding is consistent with the consensus in the literature about the advantages of forecast combination (see Stock and Watson, 1998, Timmermann, 2006).
Table 6

MAE across sub-samples.

MAE
1st sub-sample
2nd sub-sample
1 week2 weeks3 weeks4 weeks1 week2 weeks3 weeks4 weeks
CO130419792486304560137347974013744
UM10631330170025595681639973548640
UA1544248241185870620188211179014907
GT114414301847251250096846901311601
MO10341957322150264516530061827797
PS1671282743576367487970361141416466
LA158330034983704842605739827512443
JH13721431331870168879104531200413382
EN9451168153518994943627475189422
CE116913401636220453656752831810137
PO65418303848654124155515989615132

Relative MAE

1st sub-sample
2nd sub-sample
1 week2 weeks3 weeks4 weeks1 week2 weeks3 weeks4 weeks

CO1.991.080.650.472.491.330.980.91
UM1.630.730.440.392.351.160.740.57
UA2.361.361.070.902.571.601.190.99
GT1.750.780.480.382.071.240.910.77
MO1.581.070.840.771.870.960.620.52
PS2.561.541.130.972.021.281.151.09
LA2.421.641.291.081.761.040.840.82
JH2.100.780.861.073.681.901.210.88
EN1.440.640.400.292.051.140.760.62
CE1.790.730.430.342.221.220.840.67
PO1.001.001.001.001.001.001.001.00

Notes: The table reports the MAE of forecast errors for each team, the ensemble (EN), the core ensemble (CE), and the polynomial (PO) forecasts. The top panel shows the MAE level, and the bottom panel shows the MAE relative to the MAE of the benchmark model. The first evaluation sub-sample is from June 20, 2020 to October 31, 2020, and the second evaluation sub-sample from November 7, 2020 to March 20, 2021.

Turning to the second sub-sample from November 7, 2020 to March 20, 2021, we can see from Table 5 that the benchmark still significantly outperforms some teams and the ensemble forecasts at the shortest horizon. However, in this sub-sample, the forecasting teams and the ensembles fail to significantly outperform the benchmark even at the longer horizons. The MAEs for the second sub-sample reported in Table 6 indicate that, also in this sub-sample, at 3- and 4-week horizons, the two ensembles performed better than most of the individual teams. Finally, notice that the performance of the ensemble forecast published by the CDC is similar to the one of the core ensemble constructed by combining all the forecasts of the teams in Table 1. However, the MAEs for the ensemble are in all cases smaller than the ones for the core ensemble, indicating that combining a larger set of forecasts than the ones considered in Table 1 can provide some benefits in terms of predictive ability, albeit small.6 Results are similar overall, regardless of the type of estimator of the long-run variance. We also notice that findings from the bootstrap are largely the same, and confirm that fixed-smoothing asymptotics is a suitable and computationally much less time-consuming alternative to bootstrapping, as also found in Coroneo and Iacone (2020) and Gonçalves and Vogelsang (2011). Moreover, by using fixed-smoothing asymptotics, we have known critical values for each test, given the sample size and choice of bandwidth.

Alternative loss functions

The absolute value loss function that we use in the baseline forecast evaluation reported in Table 5 is a common choice in forecast evaluation. In particular, the null hypothesis is the equality of the mean absolute prediction error. However, in relation to predicting the spread of COVID-19 (and, more generally, predicting the spread of an epidemic), the cost of under-predicting the spread of the disease can be greater than the cost of over-predicting it. Similarly, scale effects are important, since the same forecast error may be more costly for public health policy interventions when the number of fatalities is small compared to when it is large. For these reasons, in this section we consider alternative loss functions.7 The DM test can be applied directly to alternative loss functions. Thus, we consider three alternative loss functions that provide alternative criteria for forecast comparison. Denoting as the forecast error (thus abbreviating in this way, ), the alternative loss functions considered are the following: Quadratic: ; Absolute percentage: ; Linex: . The quadratic loss function is a popular measure whereby large forecast errors are penalized more: in this case, it seems natural to interpret it as giving more weight to fatalities that happen when the epidemic is less predictable. The absolute percentage loss considers the scale of the number of fatalities occurring in the period, thus allowing for different evaluations of the same forecast error when only a few fatalities occur, as opposed to when there are a large number of fatalities. Finally, with the linear exponential (linex) loss function, we impose asymmetric weights, with more penalty given to under-prediction than to over-prediction. This reflects the fact that the social cost of the two errors, under- and over-prediction, are different, as the cost of not responding to the pandemic and incurring in a large loss of lives in the future is often regarded to be much higher than the economic and social cost of responding too quickly, by imposing a lockdown when it is not necessary (on the precautionary principle in public health, see Goldstein, 2001). The findings are summarized in Fig. 4, Fig. 5 for the first evaluation sub-sample, and in Fig. 6, Fig. 7 for the second evaluation sub-sample. The results for the absolute value loss function are also included, to facilitate the comparison. The dotted, dashed, and continuous red horizontal lines respectively denote the 20%, 10%, and 5% significance levels, which are, respectively, , , and for fixed- asymptotics and , , and for fixed- asymptotics.
Fig. 4

Forecast evaluation with WCE – first evaluation sub-sample.

This figure reports the test statistic for the test of equal predictive accuracy using the weighted covariance estimator (WCE) of the long-run variance and fixed- asymptotics. The benchmark is a second-degree polynomial model fitted on a rolling window of five observations. A positive value of the test statistic indicates lower loss for the forecaster, i.e. better performance of the forecaster relative to the polynomial model. Different loss functions are reported with different markers: a plus sign refers to a quadratic loss function, a diamond to the absolute loss function, a filled circle to the absolute percentage loss function, and an empty circle to the asymmetric loss function. The dotted, dashed, and continuous red horizontal lines respectively denote the 20%, 10%, and 5% significance levels. The forecast horizons are 1, 2, 3, and 4 weeks ahead. The evaluation sample is from June 20, 2020 to October 31, 2020.

Fig. 5

Forecast evaluation with WPE – first evaluation sub-sample.

This figure reports the test statistic for the test of equal predictive accuracy using the weighted periodogram estimator (WPE) of the long-run variance and fixed- asymptotics. The benchmark is a second-degree polynomial model fitted on a rolling window of five observations. A positive value of the test statistic indicates lower loss for the forecaster, i.e. better performance of the forecaster relative to the polynomial model. Different loss functions are reported with different markers: a plus sign refers to a quadratic loss function, a diamond to the absolute loss function, a filled circle to the absolute percentage loss function, and an empty circle to the asymmetric loss function. The dotted, dashed, and continuous red horizontal lines respectively denote the 20%, 10%, and 5% significance levels. The forecast horizons are 1, 2, 3, and 4 weeks ahead. The evaluation sample is from June 20, 2020 to October 31, 2020.

Fig. 6

Forecast evaluation WCE – second evaluation sub-sample.

This figure reports the test statistic for the test of equal predictive accuracy using the weighted covariance estimator (WCE) of the long-run variance and fixed- asymptotics. The benchmark is a second-degree polynomial model fitted on a rolling window of five observations. A positive value of the test statistic indicates lower loss for the forecaster, i.e. better performance of the forecaster relative to the polynomial model. Different loss functions are reported with different markers: a plus sign refers to a quadratic loss function, a diamond to the absolute loss function, a filled circle to the absolute percentage loss function, and an empty circle to the asymmetric loss function. The dotted, dashed, and continuous red horizontal lines respectively denote the 20%, 10%, and 5% significance levels. The forecast horizons are 1, 2, 3, and 4 weeks ahead. The evaluation sample is from November 7, 2020 to March 20, 2021.

Fig. 7

Forecast evaluation WPE - second evaluation sub-sample.

This figure reports the test statistic for the test of equal predictive accuracy using the weighted periodogram estimator (WPE) of the long-run variance and fixed- asymptotics. The benchmark is a second-degree polynomial model fitted on a rolling window of five observations. A positive value of the test statistic indicates lower loss for the forecaster, i.e. better performance of the forecaster relative to the polynomial model. Different loss functions are reported with different markers: a plus sign refers to a quadratic loss function, a diamond to the absolute loss function, a filled circle to the absolute percentage loss function, and an empty circle to the asymmetric loss function. The dotted, dashed, and continuous red horizontal lines respectively denote the 20%, 10%, and 5% significance levels. The forecast horizons are 1, 2, 3, and 4 weeks ahead. The evaluation sample is from November 7, 2020 to March 20, 2021.

Forecast evaluation with WCE – first evaluation sub-sample. This figure reports the test statistic for the test of equal predictive accuracy using the weighted covariance estimator (WCE) of the long-run variance and fixed- asymptotics. The benchmark is a second-degree polynomial model fitted on a rolling window of five observations. A positive value of the test statistic indicates lower loss for the forecaster, i.e. better performance of the forecaster relative to the polynomial model. Different loss functions are reported with different markers: a plus sign refers to a quadratic loss function, a diamond to the absolute loss function, a filled circle to the absolute percentage loss function, and an empty circle to the asymmetric loss function. The dotted, dashed, and continuous red horizontal lines respectively denote the 20%, 10%, and 5% significance levels. The forecast horizons are 1, 2, 3, and 4 weeks ahead. The evaluation sample is from June 20, 2020 to October 31, 2020. Forecast evaluation with WPE – first evaluation sub-sample. This figure reports the test statistic for the test of equal predictive accuracy using the weighted periodogram estimator (WPE) of the long-run variance and fixed- asymptotics. The benchmark is a second-degree polynomial model fitted on a rolling window of five observations. A positive value of the test statistic indicates lower loss for the forecaster, i.e. better performance of the forecaster relative to the polynomial model. Different loss functions are reported with different markers: a plus sign refers to a quadratic loss function, a diamond to the absolute loss function, a filled circle to the absolute percentage loss function, and an empty circle to the asymmetric loss function. The dotted, dashed, and continuous red horizontal lines respectively denote the 20%, 10%, and 5% significance levels. The forecast horizons are 1, 2, 3, and 4 weeks ahead. The evaluation sample is from June 20, 2020 to October 31, 2020. Forecast evaluation WCE – second evaluation sub-sample. This figure reports the test statistic for the test of equal predictive accuracy using the weighted covariance estimator (WCE) of the long-run variance and fixed- asymptotics. The benchmark is a second-degree polynomial model fitted on a rolling window of five observations. A positive value of the test statistic indicates lower loss for the forecaster, i.e. better performance of the forecaster relative to the polynomial model. Different loss functions are reported with different markers: a plus sign refers to a quadratic loss function, a diamond to the absolute loss function, a filled circle to the absolute percentage loss function, and an empty circle to the asymmetric loss function. The dotted, dashed, and continuous red horizontal lines respectively denote the 20%, 10%, and 5% significance levels. The forecast horizons are 1, 2, 3, and 4 weeks ahead. The evaluation sample is from November 7, 2020 to March 20, 2021. Forecast evaluation WPE - second evaluation sub-sample. This figure reports the test statistic for the test of equal predictive accuracy using the weighted periodogram estimator (WPE) of the long-run variance and fixed- asymptotics. The benchmark is a second-degree polynomial model fitted on a rolling window of five observations. A positive value of the test statistic indicates lower loss for the forecaster, i.e. better performance of the forecaster relative to the polynomial model. Different loss functions are reported with different markers: a plus sign refers to a quadratic loss function, a diamond to the absolute loss function, a filled circle to the absolute percentage loss function, and an empty circle to the asymmetric loss function. The dotted, dashed, and continuous red horizontal lines respectively denote the 20%, 10%, and 5% significance levels. The forecast horizons are 1, 2, 3, and 4 weeks ahead. The evaluation sample is from November 7, 2020 to March 20, 2021. First of all, we can observe how the results are not substantially different whether adopting a WCE or WPE estimator, as already documented in Table 5. Fig. 4, Fig. 5, Fig. 6, Fig. 7 also show that changing the loss function to quadratic or absolute percentage does not have great impact on the evaluation of predictive ability. On the other hand, the results are different if the linex function is used: in this case, for forecast horizons larger than 1 week ahead, the null hypothesis is almost never rejected at the 5% significance level. Considering the forecast horizon, it is clear that the simple polynomial benchmark outperforms all the teams, sometimes significantly so, at the 1-week horizon, and often at the 2-week horizon. However, as the forecasting horizon moves to 3 and 4 weeks, the teams improve their performance with respect to the polynomial benchmark. In the first evaluation sub-sample, the Georgia Institute of Technology, Deep Outbreak Project (GT), and the University of Massachusetts, Amherst (UM) teams, and also the CDC ensemble and the core ensemble, outperform the benchmark at almost any level of statistical significance when the quadratic and the absolute percentage loss functions are used. In the second evaluation sub-sample, when we use the quadratic and the absolute percentage loss functions, we still document more accurate predictions for several teams, for example for the University of Massachusetts, Amherst (UM), for Northeastern University, Laboratory for the Modeling of Biological and Socio-technical Systems (MO), and for the ensembles, although these findings are seldom statistically significant. On the other hand, neither the forecasting teams nor the ensemble forecasts outperform the benchmark significantly when the linex loss function is used. This seems to be mainly due to the fact that most forecasting teams (and the ensembles) under-predicted the fatalities, and this is more penalized with this loss function. Our empirical findings may also be viewed as offering support for the results discussed by Elliott and Timmermann (2004), who show how the equal-weights ensemble is less appropriate in the presence of asymmetries in the loss function and in the distribution of the errors. Notice, however, that the teams submitting forecasts to the CDC were advised that their point forecasts would be evaluated with the absolute value loss function, and therefore it is fair to conjecture that their predictions do not optimize the quadratic, the absolute percentage, or the linex loss functions. Had the teams been told that an alternative loss function was to be used to evaluate the forecast accuracy, then their point predictions might have been different. For example, if the linex loss function were used, the predictions would in general be larger. In general, we conclude that the ensemble forecasts deliver some of the best performing predictions. They often achieve statistically significant outperformance against the benchmark. This is the case for the 3- and 4-week-ahead predictions during the first evaluation sub-sample, when losses are evaluated using the absolute value, quadratic, or absolute percentage loss function. Even when the outperformance is not significant, ensemble predictions perform relatively well, in the sense of not underperforming the benchmark, even during the second evaluation sub-sample when the forecast errors are larger. Between the two ensemble forecasts, the wider ensemble obtained by the CDC performs slightly better against the benchmark compared to the core ensemble, illustrating once again the gains from combining a large number of predictions.

Comparing the two evaluation sub-samples

Comparing the results of the tests for equal predictive ability across the two sub-samples, we note that the null hypothesis is more difficult to reject during the second evaluation sub-sample. In particular, whereas during the first sub-sample (Fig. 4, Fig. 5), many forecasting teams and the ensemble forecasts outperform the benchmark at the 3- and 4-week horizons, this is no longer true in the second sub-sample (Fig. 6, Fig. 7). Heightened incident deaths, associated with the increased transmissibility of the new strains of the virus that emerged in late 2020 and in 2021, may have affected the statistical performance of the tests of equal predictive ability in two ways: by making the task of forecasting more difficult (as evident by the larger MAEs in the second sub-sample in Table 6), and by inflating the long-run variance of the test statistic (thus reducing the power to detect a significant difference for given MAE differential). Of course, the purpose of the tests of equal predictive ability is not to compare predictive ability across different periods, so a more pronounced failure to reject the null hypothesis during the second evaluation sub-sample is not evidence that the models were less valuable in this period. To examine further the reasons for the apparent change in the forecastability of the epidemic, in the bottom panel of Table 6 we report the ratio of the MAE of each forecasting team and the MAE of the benchmark. This enables us to compare the relative performance of each forecasting team and the benchmark across the two evaluation sub-samples. Considering for example the 4-week forecasting horizon, we can notice that for some forecasting teams and the ensembles, the ratios are considerably smaller for the first evaluation sub-sample compared to the second evaluation sub-sample. This is true for all the loss functions considered (MAE, RMSE, MAPE, and MLinex), and suggests that the forecasting environment was different across the first and the second evaluation sub-samples.8

Additional experiments

In the Online Appendix, we consider several additional experiments and empirical exercises. In particular, in Appendix C, we perform the tests of equal predictive ability for the full sample, instead of considering each evaluation sub-sample separately. In Appendix D, we use an alternative benchmark model, obtained based on fitting an AR(1) model to the log incidence of weekly deaths. Finally, in Appendix E, we use the CDC ensemble forecast as the benchmark model, to test formally the null of equal predictive ability of the forecasting teams and the ensemble forecast.

Conclusion

We evaluated the relative predictive accuracy of real-time forecasts for the number of COVID-19 fatalities in the US produced by several competing forecasting teams for two evaluation periods: June 20, 2020 to October 31, 2020 (first evaluation sub-sample); and November 7, 2020 to March 20, 2021 (second evaluation sub-sample). Ensemble forecasts that combine all available forecasts using an equal-weights scheme were also included. Since the sample sizes are small, we used fixed-smoothing asymptotics for the limit distribution of the scaled expected loss differential between two competing forecasts. We found that none of the forecasting teams outperformed a simple statistical benchmark at the 1-week horizon; however, at longer forecasting horizons some teams showed superior predictive ability. The ensemble forecasts delivered some of the most competitive predictions. Whilst they did not yield the best forecasts overall, they were competitive in the sense of delivering predictions that significantly outperformed the benchmark at longer horizons during the first evaluation sub-sample, and they never performed statistically worse than the benchmark, even in the second evaluation sub-sample. In this sense, the ensemble forecast may be seen as a robust choice. We also documented that the broader ensemble published by the CDC was more accurate than the core ensemble that only pooled forecasts from the teams that we included in our exercise. Overall, our results indicate that forecasts of the COVID-19 epidemic are valuable but need to be used with caution, and decision-makers should not rely on a single forecasting team (or a small set) to predict the evolution of the pandemic, but should hold a large and diverse portfolio of forecasts. A natural extension of our analysis is to evaluate the interval forecasts with different levels of coverage submitted by the forecasting teams to the CDC. This would require choosing an appropriate loss function, for example the weighted interval score (see Bracher, Ray, Gneiting, & Reich, 2021), and applying the same alternative asymptotics we used here. In particular, recent work by Coroneo, Iacone, and Profumo (2019) shows that fixed-smoothing asymptotics may also be employed successfully to evaluate density forecasts. Another interesting extension to the current work is to consider the predictive accuracy for a panel data of forecasts (since the forecasting teams predict not only the national spread of the disease but also the regional evolution of the epidemic). Timmermann and Zhu (2019) propose methods for testing predictive accuracy for panel forecasts. In particular, they develop a panel-data Diebold–Mariano test for equal predictive accuracy that pools information across both the time-series and cross-sectional dimensions. Our analysis could, therefore, be extended to the evaluation of a panel of regional predictions.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  7 in total

Review 1.  The precautionary principle also applies to public health actions.

Authors:  B D Goldstein
Journal:  Am J Public Health       Date:  2001-09       Impact factor: 9.308

2.  Real-time forecasting of epidemic trajectories using computational dynamic ensembles.

Authors:  G Chowell; R Luo; K Sun; K Roosa; A Tariq; C Viboud
Journal:  Epidemics       Date:  2019-12-21       Impact factor: 4.396

3.  When will the Covid-19 pandemic peak?

Authors:  Shaoran Li; Oliver Linton
Journal:  J Econom       Date:  2020-09-08       Impact factor: 2.388

4.  Evaluating epidemic forecasts in an interval format.

Authors:  Johannes Bracher; Evan L Ray; Tilmann Gneiting; Nicholas G Reich
Journal:  PLoS Comput Biol       Date:  2021-02-12       Impact factor: 4.779

5.  Accuracy of real-time multi-model ensemble forecasts for seasonal influenza in the U.S.

Authors:  Nicholas G Reich; Craig J McGowan; Teresa K Yamana; Abhinav Tushar; Evan L Ray; Dave Osthus; Sasikiran Kandula; Logan C Brooks; Willow Crawford-Crudell; Graham Casey Gibson; Evan Moore; Rebecca Silva; Matthew Biggerstaff; Michael A Johansson; Roni Rosenfeld; Jeffrey Shaman
Journal:  PLoS Comput Biol       Date:  2019-11-22       Impact factor: 4.475

6.  Emergence of SARS-CoV-2 B.1.1.7 Lineage - United States, December 29, 2020-January 12, 2021.

Authors:  Summer E Galloway; Prabasaj Paul; Duncan R MacCannell; Michael A Johansson; John T Brooks; Adam MacNeil; Rachel B Slayton; Suxiang Tong; Benjamin J Silk; Gregory L Armstrong; Matthew Biggerstaff; Vivien G Dugan
Journal:  MMWR Morb Mortal Wkly Rep       Date:  2021-01-22       Impact factor: 17.586

  7 in total
  1 in total

1.  Short-term Covid-19 forecast for latecomers.

Authors:  Marcelo C Medeiros; Alexandre Street; Davi Valladão; Gabriel Vasconcelos; Eduardo Zilberman
Journal:  Int J Forecast       Date:  2021-10-13
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.