Literature DB >> 35013654

A random forest model for forecasting regional COVID-19 cases utilizing reproduction number estimates and demographic data.

Joseph Galasso1, Duy M Cao2, Robert Hochberg3.   

Abstract

During the COVID-19 pandemic, predicting case spikes at the local level is important for a precise, targeted public health response and is generally done with compartmental models. The performance of compartmental models is highly dependent on the accuracy of their assumptions about disease dynamics within a population; thus, such models are susceptible to human error, unexpected events, or unknown characteristics of a novel infectious agent like COVID-19. We present a relatively non-parametric random forest model that forecasts the number of COVID-19 cases at the U.S. county level. Its most prioritized training features are derived from easily accessible, standard epidemiological data (i.e., regional test positivity rate) and the effective reproduction number ( R t ) from compartmental models. A novel input training feature is case projections generated by aligning estimated effective reproduction number (pre-computed by COVIDActNow.org) with real time testing data until maximally correlated, helping our model fit better to the epidemic's trajectory as ascertained by traditional models. Poor reliability of R t is partially mitigated with dynamic population mobility and prevalence and mortality of non-COVID-19 diseases to gauge population disease susceptibility. The model was used to generate forecasts for 1, 2, 3, and 4 weeks into the future for each reference week within 11/01/2020 - 01/10/2021 for 3068 counties. Over this time period, it maintained a mean absolute error (MAE) of less than 300 weekly cases/100,000 and consistently outperformed or performed comparably with gold-standard compartmental models. Furthermore, it holds great potential in ensemble modeling due to its potential for a more expansive training feature set while maintaining good performance and limited resource utilization.
© 2022 Elsevier Ltd. All rights reserved.

Entities:  

Keywords:  COVID-19; Compartmental model; Mobility; Random forest; US county

Year:  2022        PMID: 35013654      PMCID: PMC8731233          DOI: 10.1016/j.chaos.2021.111779

Source DB:  PubMed          Journal:  Chaos Solitons Fractals        ISSN: 0960-0779            Impact factor:   5.944


Introduction

The COVID-19 epidemic in the United States proved devastating economically as it is projected to cause 3.2 to 4.8 trillion USD in net U.S. GDP loss [1]. The epidemic has had a devastating toll on life, particularly among the elderly and members of ethnic minorities such as African Americans [2]. Throughout the COVID-19 pandemic, it has been necessary to forecast the progression of the pandemic at the U.S. county/county-equivalent (CCE) level so that local authorities can effectively implement public health measures such as social distancing or quarantines [3]. This need is particularly great due to findings that there is low reliability between state-wide and county-specific reported data, requiring that the pandemic is tracked at the most granular level possible [4]. Forecasts have been traditionally done with compartmental models such as Susceptible-Exposed-Infectious-Recovered (SEIR) [5], [6]. These models segment the population into various compartments for stages in progression of the disease of interest. Transitions between these compartments represent epidemic dynamics. These models can be used to solve for the time-path of the , which is the estimated ratio of new infections caused by each currently infected individual, over the course of an epidemic; this in turn can be used to forecast future epidemic progression [6]. However, these models involve numerous assumptions in their design about disease spread dynamics and their interpretability and usability is limited by the quality of these assumptions [5], [6], [7]. Thus, relatively non-parametric deep or machine-learning models, such as Random Forests (RFs), are attractive alternatives to compartmental models, as they avoid assumptions about the distribution of input data and generate forecasts based on observed empirical trends in this data [8], [9]. In addition to being non-parametric, RF-regressors are highly effective at extracting non-linear relationships from input data while being time efficient [10], [11]. First described by Breiman, RF-regressors are ensemble models of regression trees each trained on different subsets of input data, reducing the variance of predictors and minimizing overfitting [12]. In addition, RFs enable relatively easy estimations of variable importance [13], which can serve as an assessment of model performance. RFs have been successfully utilized in prior studies for predicting diarrheal infections, Dengue, H5N1 influenza, and West Nile virus [10], [14], [15], [16], [17]. They also have been applied for short-term forecasting of COVID-19 infections by Ribeiro et al. [11]. We similarly apply an RF to project COVID-19 infections at the more granular CCE level up to 4 weeks in the future. Unlike prior RF epidemiology forecasting models, we use a unique and extensive combination of features including population health and mobility, demographic, and COVID-19 testing data. We develop a novel method of computing case and forecasts with an algorithm that is based on our observation that COVID-19 cases and SEIR-generated have a similar trajectory over the course of an epidemic, but lags behind cases, making it predictive of cases in the lag period. These forecasts are then added as a training feature for our RF model, which produces the final forecasts.

Methods

Data acquisition

Our time-series datasets begin on 03/31/2020 for 3068 U.S. CCEs with complete, consistent data that could be processed without error in our downstream pipelines. To normalize testing and case counts by population, these metrics were converted to incidence rate (IR) using the 2018 U.S. Census CCE-level population estimates in Eq. (1).

Generation of and case alignment forecast features

For any given date in every U.S. CCE, we forecast and testing-normalized cases for 1, 2, 3, and 4 weeks into the future. First, a testing-normalized COVID-19 cases time-series was generated by dividing the “Weekly case increase” feature by the “Daily tests increase” feature (see Table 1 ). The normalized cases time-series and the “Daily estimated ” feature were used to generate and case-prediction features, as shown in Fig. 1 . All linear regression models were implemented with the module [28].
Table 1

Raw training data sources and normalizations. Description of datasets, variables extracted, regional level, and applied normalization in the training dataset.

Raw dataset sourceFeature(s)Geographic levelTransformation applied
Johns Hopkins University (JHU) CSSE [18]Weekly case increaseCCE-levelRolling 7-day sum of case IR (Eq. 1)
Facebook.com [19]Daily mobility relative to average baseline, proportion of users staying in same locationCCE-levelRolling 7-day mean
COVID Tracking Project [20]Daily tests increase, test positivityState-levelRolling 7-day mean of test IR (Eq. 1) and test positivity
COVIDActNow.org [21]Daily estimated RtCCE-level and state-levelNone
Surgo Foundation [37]Metric that assesses CCE vulnerability to COVID-19, taking into account socioeconomic, epidemiological, and heatlhcare system risk factorsCCE-levelNone
Institute for Health Metrics and Evaluation (IHME) [22]Infectious disease mortality rates (tuberculosis, AIDS, diarrheal disease, lower respiratory disease, meningitis, hepatitis)CCE-levelNone
IHME [23]Respiratory disease mortality rates (interstitial lung disease, asthma, coal pneumoconiosis, asbestosis, silicosis, pneumoconiosis, COPD, chronic respiratory disease, other pneumoconiosis, other respiratory diseases)CCE-levelNone
IHME [24]Mortality risk (0–5, 5–25, 25–45, 45–65, and 65–85 age groups), life expectancyCCE-levelNone
IHME [25]Diabetes prevalence ratesCCE-levelNone
IHME [26]Obesity prevalence rates (combined male and female)CCE-levelNone
U.S. Census (2018 estimates) [27]Prevalence of African Americans, Native Americans, Hispanic Americans, Multiracial Americans, and individuals over 65 years of ageCCE-levelNone
Fig. 1

and case prediction feature generation for a CCE. This procedure is repeated for 7-, 14-, 21- and 28-day forecasts. In Fig. 1A and Fig. 1B, for the CCE and its state are both separately considered; whichever achieves the highest Pearson correlation of any forward shift x that is <50 days (i.e. optimal shift) is used for the regression model in Fig. 1B. The extrapolation in Fig. 1E is calculated by linear regression models trained on the last 14 defined values of each curve; curves are extrapolated to the target end date (i.e. 7, 14, 21, or 28 days in the future). For Fig. 1F, curves in prediction time have forecast values relative to real time; thus, for 28-day forecasts, values are those forecast 28 days into the future.

Raw training data sources and normalizations. Description of datasets, variables extracted, regional level, and applied normalization in the training dataset. and case prediction feature generation for a CCE. This procedure is repeated for 7-, 14-, 21- and 28-day forecasts. In Fig. 1A and Fig. 1B, for the CCE and its state are both separately considered; whichever achieves the highest Pearson correlation of any forward shift x that is <50 days (i.e. optimal shift) is used for the regression model in Fig. 1B. The extrapolation in Fig. 1E is calculated by linear regression models trained on the last 14 defined values of each curve; curves are extrapolated to the target end date (i.e. 7, 14, 21, or 28 days in the future). For Fig. 1F, curves in prediction time have forecast values relative to real time; thus, for 28-day forecasts, values are those forecast 28 days into the future.

RF training and forecasting

The dataset used to train and validate the RF forecasting models includes all features in Table 1 for each U.S. CCE; however, “Weekly case increase” and “Daily estimated ” features were replaced with the Fig. 1F features for 7-, 14-, 21-, and 28-day forecasts. Thus, the preliminary forecasts were used as features to generate the final forecasts. All features were normalized by removing their mean and scaling to unit variance. For each Sunday from 11/01/2020 to 01/10/2021, we trained and validated RF forecasting models via incremental learning. Thus, we filtered out feature data that occurs on or after the Sunday of interest, but later Sundays will have more feature data than earlier Sundays. Then, we randomly split the feature data for a given Sunday into a training subset and a validation subset with a 9:1 ratio, respectively. The training subset is used to train a RF regression model for each forecast timepoint (7, 14, 21, or 28 days into the future); the target outcome was the “Weekly case increase” from Table 1 shifted backwards to represent future cases on the forecast timepoint. The validation subset is used to ensure that trained RF models do not overfit the training subset. Finally, data for the Sunday of interest is input to the trained RF models to generate forecasts for each forecast timepoint. RF regression models are ensemble machine learning algorithms that were first described by Breiman [12]. They create multiple regression trees trained on unique bootstrap samples of the training dataset with a random subset of the input features. The output of all trees is averaged to create the final projection. We used the Scikit-learn (version 0.23.2) implementation with default hyperparameters and 20 estimators [28].

Model evaluation

The permutation importance of all features input into the RF models was calculated as described by Breiman as the decrease in mean squared error of the model when a feature of interest is randomly shuffled [12]. We used two metrics, MAE and R-squared (), to evaluate the accuracy of our model’s forecasts relative to actual case counts for both the training dataset and forecasts outside of the training dataset. These metrics are calculated as follows: We also used these metrics to calculate the error for equivalent forecasts by the JHU Infectious Disease Dynamics group (IDD), the JHU Applied Physics Lab (APL), and One Quiet Night (OQN) forecasting models [29], [30], [31]. These were selected as they are parts of the U.S. Centers for Disease Control and Prevention (U.S. CDC) ensemble model, have relatively high performance, and forecast for a large breadth (2349) of U.S. counties along with the RF model [32].

Results and discussion

Analysis of and case alignment forecasts

When the case time-series for a CCE was normalized by population and state-level testing data, it often showed a very similar shape to the CCE and/or state time-series, as shown in Fig. 1. This should be expected, as both time-series indicate new infection load over the course of an epidemic, although lags behind cases, which we suspect is because represents infection load in the present moment, but these infections will not be detected via testing until much later due to viral incubation periods of 2–14 days and test result reporting delays [33]. In the 01/10/2021 dataset, the average optimal shift for the most highly correlated time-series, whether state or CCE, was 34 days forward. However, there were many CCEs where this relationship was weak. We observed that 1782 CCEs selected, optimally shifted time-series have a Pearson correlation with the case time-series . They also have a mean population density of 89 people/mi, vs. the 273 people/mi mean over the entire set of 3068 CCEs. Thus, we attribute the low correlations to poor and/or inconsistent testing efforts and data quality in rural CCEs. Further supporting this is our observation that in CCEs where we select their state’s time-series as opposed to their own, the mean population density is also relatively low at 210 people/mi. In concurrence with Omori et al., we found that normalization of the case time-series with testing data is critical to expose underlying changes in COVID-19 progression, as seen in Fig. 2 [34]. However, our approach is limited by use of state-level testing data as opposed to CCE-level testing data, which was inaccessible to us. However, state-level testing data has the advantage of including individuals who are not tested in their CCE of residence due to inequity in regional testing access.
Fig. 2

time-series lag behind case time-series used to forecast cases in lag period. In Harris County, TX, when the CCE time-series from 05/22/2020 to 07/10/2020 has a maximum Pearson correlation with the CCE’s testing and population normalized case time-series for the same period when shifted forward 10 days (B); also, this correlation is higher than that obtained by any shift of the state time-series. Thus, the shifted CCE time-series is linearly regressed against cases (C). This model is applied to unshifted CCE time-series to generate a case-prediction time-series; the last 10 days of both these time-series are predictive of the next 10 days of cases beyond 07/10/2020.

time-series lag behind case time-series used to forecast cases in lag period. In Harris County, TX, when the CCE time-series from 05/22/2020 to 07/10/2020 has a maximum Pearson correlation with the CCE’s testing and population normalized case time-series for the same period when shifted forward 10 days (B); also, this correlation is higher than that obtained by any shift of the state time-series. Thus, the shifted CCE time-series is linearly regressed against cases (C). This model is applied to unshifted CCE time-series to generate a case-prediction time-series; the last 10 days of both these time-series are predictive of the next 10 days of cases beyond 07/10/2020.

RF feature prioritization and training performance

On average over the 11 epi weeks, RFs always prioritized population-normalized, state-level test numbers, test positivity, and case and alignment forecasts as seen in Fig. 3 . This validates the model, as these features most directly measure COVID-19 progression. Importantly, of all 8 alignment forecasts, those for the desired forecast target were always prioritized highest. Conversely, demographic, population health, and population mobility features are relatively inconsequential; on average across Fig. 3A-D, these 46 features share of the total sum of all 56 features permutation scores is just 14.33%.
Fig. 3

Top 7 RF features for each forecast target. For each forecast target, RF feature permutation importances were averaged over all 11 epi weeks and the top 7 features are shown in the subfigures above along with their standard deviation as error bars.

Top 7 RF features for each forecast target. For each forecast target, RF feature permutation importances were averaged over all 11 epi weeks and the top 7 features are shown in the subfigures above along with their standard deviation as error bars. On the training datasets for all 4 forecast targets and all 11 epi weeks, was very stable, averaging 0.97 with a standard deviation of 0.00. For the corresponding validation datasets, fell to 0.92 with a standard deviation of 0.02. Thus, overfitting is not extreme and, considering the RFs relatively high MAE and on actual case data in comparison to other models as seen in Figs. 4 and 5 , is not detrimental to our objectives.
Fig. 4

Performance evaluation using MAE. The errors between projections and real number of incident cases were calculated using Eq. (3). The y-axes of the graphs have been limited so that all models can be visually compared.

Fig. 5

Performance evaluation using . As with Fig. 4, the y-axes of the graphs have been limited. The proportions of variance between projections and observed values were evaluated using Eq. (2). We notice that there are large anomalies in the weekly of the Google_Harvard and JHUAPL_Bucky models after epi week 202051; however, for the sake of complete comparison, all weeks for all models are shown.

Performance evaluation using MAE. The errors between projections and real number of incident cases were calculated using Eq. (3). The y-axes of the graphs have been limited so that all models can be visually compared. Performance evaluation using . As with Fig. 4, the y-axes of the graphs have been limited. The proportions of variance between projections and observed values were evaluated using Eq. (2). We notice that there are large anomalies in the weekly of the Google_Harvard and JHUAPL_Bucky models after epi week 202051; however, for the sake of complete comparison, all weeks for all models are shown.

RF forecasting performance vs. other models

As may be observed in Figs. 4 and 5, over the 11 reference weeks for which forecasts were generated by the RF models, MAE and remain relatively consistent compared to the OQN and JHU models. The RF’s is competitive with and often higher than the for other models, notably November 2020. Cases in November were harder to model accurately, as November saw a 40% increase in cases in the 3068 counties in its fourth week relative to its first week, whereas December only saw an 8% decrease in cases for its equivalent comparison period. Periods of relatively low performance are generally shared by all models, indicating that case load changes during these weeks are simply less predictable. For each epi week, case projections from each model were collected and compared to actual cases that occurred during these weeks as reported by JHU [18]. The results were calculated and visualized in Figs. 4 and 5. The JHU IDD and APL models are SEIR models [29], [30], whereas the OQN model applies a linear regression model to each CCE [31]. The distribution of output from a model exclusive to a CCE will be skewed towards the distribution of its training dataset labels, which may be a factor explaining OQN’s low MAE. On the other hand, the RF model forecasts for all counties, affording it a larger dataset which possibly contributes to its relatively high, stable fit to the actual cases as indicated by its .

Conclusion

We present a unique method to project COVID-19 cases for CCEs by using their or their state’s time-series to predict cases, taking advantage of the backward lag of regional time-series from the case time-series despite their similar trajectory. These predictions are input into a RF regression models with regional testing, demographic, population mobility and population health data for final case forecasts. Our approach is computationally inexpensive while remaining very effective, as our model achieves consistently high and low MAE relative to gold standard models used in the U.S. CDC ensemble model during a highly dynamic case spike period in November 2020 early January 2021. This model is limited to forecasting cases detected by testing as opposed to latent, asymptomatic cases, which can be estimated by compartmental models [35]. Thus, its best use scenario is for public health officials to identify potential outbreaks in their community to help them optimize their response. It shows evidence of good consistency in its current iteration but has some potential for improvement via addition of new features to its training dataset, particularly case and forecasts from other compartmental models. Such ensemble forecasting approaches have improved model performance significantly [36].

Availability of software and other materials

All code, programs, and data supporting the conclusions made in this article are publicly available on GitHub, located at https://github.com/solveforj/pandemic-central.

CRediT authorship contribution statement

Joseph Galasso: Conceptualization, Methodology, Software, Data curation, Writing – original draft, Writing – review & editing, Visualization, Supervision. Duy M. Cao: Conceptualization, Methodology, Software, Data curation, Writing – original draft, Writing – review & editing, Visualization. Robert Hochberg: Writing – review & editing, Supervision.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  19 in total

1.  Large-scale multivariate forecasting models for Dengue - LSTM versus random forest regression.

Authors:  Elisa Mussumeci; Flávio Codeço Coelho
Journal:  Spat Spatiotemporal Epidemiol       Date:  2020-08-28

Review 2.  Random forests for genomic data analysis.

Authors:  Xi Chen; Hemant Ishwaran
Journal:  Genomics       Date:  2012-04-21       Impact factor: 5.736

3.  The Impacts of the Coronavirus on the Economy of the United States.

Authors:  Terrie Walmsley; Adam Rose; Dan Wei
Journal:  Econ Disaster Clim Chang       Date:  2020-12-10

4.  What variables are important in predicting bovine viral diarrhea virus? A random forest approach.

Authors:  Gustavo Machado; Mariana Recamonde Mendoza; Luis Gustavo Corbellini
Journal:  Vet Res       Date:  2015-07-24       Impact factor: 3.683

5.  Variation in COVID-19 outbreaks at the US state and county levels.

Authors:  W Messner; S E Payson
Journal:  Public Health       Date:  2020-08-03       Impact factor: 2.427

6.  Short-term forecasting COVID-19 cumulative confirmed cases: Perspectives for Brazil.

Authors:  Matheus Henrique Dal Molin Ribeiro; Ramon Gomes da Silva; Viviana Cocco Mariani; Leandro Dos Santos Coelho
Journal:  Chaos Solitons Fractals       Date:  2020-05-01       Impact factor: 5.944

7.  Management strategies in a SEIR-type model of COVID 19 community spread.

Authors:  Anca Rǎdulescu; Cassandra Williams; Kieran Cavanagh
Journal:  Sci Rep       Date:  2020-12-04       Impact factor: 4.379

8.  Regional forecasting of COVID-19 caseload by non-parametric regression: a VAR epidemiological model.

Authors:  Aaron C Shang; Kristen E Galow; Gary G Galow
Journal:  AIMS Public Health       Date:  2021-02-01

9.  A scenario modeling pipeline for COVID-19 emergency planning.

Authors:  Joseph C Lemaitre; Kyra H Grantz; Joshua Kaminsky; Hannah R Meredith; Shaun A Truelove; Stephen A Lauer; Lindsay T Keegan; Sam Shah; Josh Wills; Kathryn Kaminsky; Javier Perez-Saez; Justin Lessler; Elizabeth C Lee
Journal:  Sci Rep       Date:  2021-04-06       Impact factor: 4.379

10.  Replicating and projecting the path of COVID-19 with a model-implied reproduction number.

Authors:  Shelby R Buckman; Reuven Glick; Kevin J Lansing; Nicolas Petrosky-Nadeau; Lily M Seitelman
Journal:  Infect Dis Model       Date:  2020-08-28
View more
  2 in total

1.  Design and Implementation of 3D Animation Data Processing Development Platform Based on Artificial Intelligence.

Authors:  Quansheng Gao
Journal:  Comput Intell Neurosci       Date:  2022-05-30

Review 2.  An evaluation of prospective COVID-19 modelling studies in the USA: from data to science translation.

Authors:  Kristen Nixon; Sonia Jindal; Felix Parker; Nicholas G Reich; Kimia Ghobadi; Elizabeth C Lee; Shaun Truelove; Lauren Gardner
Journal:  Lancet Digit Health       Date:  2022-10
  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.