| Literature DB >> 31712420 |
Michael A Johansson1,2, Karyn M Apfeldorf3, Scott Dobson3, Jason Devita3, Anna L Buczak4, Benjamin Baugher4, Linda J Moniz4, Thomas Bagley4, Steven M Babin4, Erhan Guven4, Teresa K Yamana5, Jeffrey Shaman5, Terry Moschou6, Nick Lothian6, Aaron Lane6, Grant Osborne6, Gao Jiang7, Logan C Brooks8, David C Farrow8, Sangwon Hyun9, Ryan J Tibshirani8,9, Roni Rosenfeld8, Justin Lessler10, Nicholas G Reich11, Derek A T Cummings12,13, Stephen A Lauer11, Sean M Moore14,15, Hannah E Clapham16, Rachel Lowe17,18, Trevor C Bailey19, Markel García-Díez20, Marilia Sá Carvalho21, Xavier Rodó18,22, Tridip Sardar22, Richard Paul23,24, Evan L Ray25, Krzysztof Sakrejda11, Alexandria C Brown11, Xi Meng11, Osonde Osoba26, Raffaele Vardavas26, David Manheim27, Melinda Moore26, Dhananjai M Rao28, Travis C Porco29, Sarah Ackley29, Fengchen Liu29, Lee Worden29, Matteo Convertino30, Yang Liu31, Abraham Reddy31, Eloy Ortiz32, Jorge Rivero32, Humberto Brito32,33, Alicia Juarrero32,34, Leah R Johnson35, Robert B Gramacy36, Jeremy M Cohen36, Erin A Mordecai37, Courtney C Murdock38,39, Jason R Rohr14,15, Sadie J Ryan13,40,41, Anna M Stewart-Ibarra42, Daniel P Weikel43, Antarpreet Jutla44, Rakibul Khan44, Marissa Poultney44, Rita R Colwell45, Brenda Rivera-García46, Christopher M Barker47, Jesse E Bell48, Matthew Biggerstaff49, David Swerdlow49, Luis Mier-Y-Teran-Romero50,10, Brett M Forshey51, Juli Trtanj52, Jason Asher53, Matt Clay53, Harold S Margolis50, Andrew M Hebbeler54,55, Dylan George55,56, Jean-Paul Chretien55,57.
Abstract
A wide range of research has promised new tools for forecasting infectious disease dynamics, but little of that research is currently being applied in practice, because tools do not address key public health needs, do not produce probabilistic forecasts, have not been evaluated on external data, or do not provide sufficient forecast skill to be useful. We developed an open collaborative forecasting challenge to assess probabilistic forecasts for seasonal epidemics of dengue, a major global public health problem. Sixteen teams used a variety of methods and data to generate forecasts for 3 epidemiological targets (peak incidence, the week of the peak, and total incidence) over 8 dengue seasons in Iquitos, Peru and San Juan, Puerto Rico. Forecast skill was highly variable across teams and targets. While numerous forecasts showed high skill for midseason situational awareness, early season skill was low, and skill was generally lowest for high incidence seasons, those for which forecasts would be most valuable. A comparison of modeling approaches revealed that average forecast skill was lower for models including biologically meaningful data and mechanisms and that both multimodel and multiteam ensemble forecasts consistently outperformed individual model forecasts. Leveraging these insights, data, and the forecasting framework will be critical to improve forecast skill and the application of forecasts in real time for epidemic preparedness and response. Moreover, key components of this project-integration with public health needs, a common forecasting framework, shared and standardized data, and open participation-can help advance infectious disease forecasting beyond dengue.Entities:
Keywords: Peru; Puerto Rico; dengue; epidemic; forecast
Mesh:
Year: 2019 PMID: 31712420 PMCID: PMC6883829 DOI: 10.1073/pnas.1909865116
Source DB: PubMed Journal: Proc Natl Acad Sci U S A ISSN: 0027-8424 Impact factor: 11.205
Fig. 1.Dengue and climate data for Iquitos, Peru and San Juan, Puerto Rico. The black and colored lines for dengue cases indicate the total and virus-specific weekly number of laboratory-confirmed cases. The yellow and red points indicate the peaks in the training and testing datasets, respectively. The climate data show the weekly rainfall (blue) and mean temperature (red) for Iquitos and San Juan, respectively, from the National Centers for Environmental Prediction Climate Forecast System Reanalysis.
Fig. 2.Weeks 12 and 24 forecasts for the 2012/2013 dengue season in Iquitos and San Juan. The solid black lines indicate the most recent data that were available to teams to inform these forecasts, and the dashed lines indicate the data that became available later in the season. The colored points represent point estimates for each team, while the bars represent 50 and 95% prediction intervals (dark and light, respectively). Forecasts for additional time points and seasons as well as for seasonal incidence are shown in , respectively.
Fig. 3.Forecast skill by team, forecast week, and target in the testing seasons (2009/2010 to 2012/2013). Solid colored lines represent the scores of individual teams averaged across all testing seasons for the respective forecast week, target, and location. For each target, the top forecast for the first 24 wk (shaded) is indicated in bold (highest average early season score). The solid black lines indicate the null model (equal probability assigned to all possible outcomes), the dashed gray lines indicate the baseline model, and the dotted black lines indicate the ensemble model. Forecasts with logarithmic scores of less than −5 are not shown. Breaks in lines indicate a score of negative infinity in at least 1 of the testing seasons.
Fig. 4.Overall forecast scores for weeks 0 to 24 in the training (2005/2006 to 2008/2009) and testing (2009/2010 to 2012/2013) seasons. Each point is the average target- and location-specific log score for a model in the training (left side; light shading) and testing (right side; dark shading) seasons. The horizontal dispersion within training and testing scores is random to improve visualization. The null forecast for each target is represented by a horizontal line. Numerous forecasts assigned 0 probability to at least 1 observed outcome. Those individual forecast probabilities were changed to 0.001 before calculating the logarithmic scores.