Literature DB >> 34249608

Spatial-temporal generalized additive model for modeling COVID-19 mortality risk in Toronto, Canada.

Cindy Feng1.   

Abstract

This article presents a spatial-temporal generalized additive model for modeling geo-referenced COVID-19 mortality data in Toronto, Canada. A range of factors and spatial-temporal terms are incorporated into the model. The non-linear and interactive effects of the neighborhood-level factors, i.e., population density and average of income, are modeled as a two-dimensional spline smoother. The change of spatial pattern over time is modeled as a three-dimensional tensor product smoother. By fitting this model, the space-time effect can uncover the underlying spatial-temporal pattern that is not explained by the covariates. The performance of the modeling method based on the individual data is also compared to the modeling methods based on the aggregated data in terms of in-sample and out-of-sample predictive checking. The results suggest that the individual-level based analysis provided a better overall model fit and higher predictive accuracy for detecting epidemic peaks in this application as compared to the analysis based on the aggregated data.
© 2021 Elsevier B.V. All rights reserved.

Entities:  

Keywords:  COVID-19 mortality; Generalized additive model; Space–time interaction; Tensor product smooth; Thin plate spline

Year:  2021        PMID: 34249608      PMCID: PMC8257405          DOI: 10.1016/j.spasta.2021.100526

Source DB:  PubMed          Journal:  Spat Stat


Introduction

Coronavirus disease (COVID-19), caused by the severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), presents an unprecedented threat to global health worldwide. Despite public health responses aimed at slowing the spread of the epidemic, confirmed case counts and hospitalizations continue to surge. To help healthcare professionals prioritize severely ill patients for providing the best possible care for patients and mitigate the burden on the healthcare system, there is an urgency to understand who is most at risk of mortality, which requires appropriate statistical approaches for the timely analysis of large datasets. For modeling COVID-19 mortality, most studies are limited to known risk factors such as older age with a high burden of co-morbid diseases (Meyerowitz-Katz and Merone, 2020, Zhou et al., 2020, Du et al., 2020). Nevertheless, COVID-19 mortality risk may also depend on the employment status, education level, income, and housing conditions (de Lusignan et al., 2020, Williamson et al., 2020), which could influence the ability to practice physical distancing measures and seek care. However, the socioeconomic factors remain understudied for modeling COVID-19 data, partly because of a lack of such information. Alternatively, neighborhood-level social factors from census data are often used as a proxy for individual-level social status. Previous research showed that neighborhoods with low socioeconomic status tend to experience higher rates of serious illness if infected due to economic and health inequities (Wadhera et al., 2020). Nevertheless, whether similar patterns have also emerged amid the COVID-19 pandemic in Canada is still understudied. Further, spatial differences in disease risk are potential indicators of space-related disease factors such as environmental exposures, spatially aggregation of infected individuals and their nonrandom social interactions or availability of sufficient health care in certain areas (Real and Biek, 2007). Quantification of heterogeneity in disease risk patterns over geographical space is, therefore, of importance. While some previous research has noted the importance of spatial autocorrelation (Mollalo et al., 2020, Cordes and Castro, 2020, Zhang and Schwartz, 2020) for modeling risk of COVID-19, the research on modeling the potential changes in spatial patterns over time is still scarce. Providing more accurate time-varying risk maps offers a perspective about which areas may be more affected and when given the time to the central decision-makers to intervene on the local policies. This study demonstrates a model developed in a generalized additive modeling (GAM) framework (Hastie and Tibshirani, 1990, Wood, 2004, Wood, 2017, Wood, 2011) is sufficiently flexible to model the spatial–temporal dynamics of COVID-19 mortality risk, which is computationally fast with good discrimination and calibration. Models with three types of link functions for the binomial regression were considered, since misspecification of the link function in binomial regression could result in substantial bias and increased mean squared error of parameter estimates and the predicted probabilities (Czado and Santner, 1992). The non-linear and interactive effects of the neighborhood-level factors, i.e., population density and the average of income, are modeled as a two-dimensional smoother. The spatial–temporal interaction is modeled as a tensor product between a two-dimensional thin-plate regression spline (TPRS) (Wood, 2003) for the centroid of the neighborhoods where the individuals are from and a one-dimensional cubic regression spline for the time variable. The spatial terms can provide insight on detecting high-risk areas not explained by the covariates and contribute to the elucidation of critical aspects of this outbreak, providing a useful perspective in the study region on how the pandemic spreads. In addition, spatial modeling of COVID-19 data often focuses on analyzing aggregated area-level data to date. The aggregated data are inexpensive to obtain, particularly through data repositories. By combining the data at individual levels at varying levels of aggregation, the size of the data could also be reduced for computational convenience. However, data aggregation may lead to ecological bias when important confounders are missing (Greenland, 2001). In this case, conclusions based on a group-level analysis may differ from those that would have been drawn had an individual-level analysis (Harbarth et al., 2001). Moreover, analyzing aggregated count data can be inherently challenging due to overdispersion, heterogeneity, or the excessive number of zeros. The choice of distributional assumptions for modeling the count outcome variable is crucial in order to draw a valid statistical inference. To date, little work has been conducted in the context of COVID-19 modeling to compare individual-level analysis versus aggregate analysis. Our study aims to fill this gap by demonstrating the potential benefits of using individual-level data compared to using the aggregated data for predictive accuracy. The predictive accuracy of the proposed model is evaluated based on both in-sample and out-of-sample predictive checking. This paper is organized as follows. In Section 2, details related to the database are provided. In Section 3, the general formulation of the spatial–temporal generalized additive model, as well as the model selection ad diagnosis methods, are presented. Section 4 presents the results obtained from the analysis. A comparison of analyses based on individual-level data and aggregated data are carried out with emphasis on predictive accuracy based on in-sample and various schemes of out-of-sample model validations. Finally, Section 5 contains the main conclusions and final remarks.

Data description

The City of Toronto in Canada contains 140 neighborhoods that were aggregated from census tracts and created by the Social Policy Analysis and Research Unit in the City’s Social Development and Administration Division with assistance from Toronto Public Health. Data on 49,216 COVID-19 confirmed cases from March 1, 2020, until December 10, 2020, in Toronto, Ontario, Canada, were retrieved from the Ontario Ministry of Health. Of those, 1938 (3.94%) died from COVID-19. The time variable provided in this dataset is a derived/combined variable that best estimates when the disease was acquired, which refers to the earliest available date from symptom onset (the first day that COVID-19 symptoms occurred), laboratory specimen collection date, or reported date. The other predictors include patient demographic (age and gender), ever hospitalized (cases that were hospitalized related to their COVID-19 infection), ever in ICU (cases that were admitted to the intensive care unit (ICU) related to their COVID-19 infection), and ever intubated (cases that were intubated related to their COVID-19 infection). Neighborhood level population density and median household income over the 140 neighborhoods in Toronto were obtained from the 2016 Canadian Census data. The daily counts of COVID-19 positive cases and mortality counts are displayed in the left panel of Fig. 1, which indicates the COVID-19 infection peaked in April–May 2020, followed by a decline in the summertime and increased substantially in Nov–Dec, 2020. However, such a pattern was not observed in the daily COVID-19 mortality count. As shown in the right panel of Fig. 1, the risk of mortality is much higher in the first wave as compared to the second wave.
Fig. 1

Left panel: Daily counts of COVID-19 positive cases (black curve) and daily mortality counts of COVID-19 (blue curve). Right panel: Daily COVID-19 case fatality rate. (For interpretation of the references to color in this figure legend, the reader is referred to the web version of this article.)

Left panel: Daily counts of COVID-19 positive cases (black curve) and daily mortality counts of COVID-19 (blue curve). Right panel: Daily COVID-19 case fatality rate. (For interpretation of the references to color in this figure legend, the reader is referred to the web version of this article.)

Methodology

Model formulation

Let denote the mortality status due to COVID-19 for the th individual, , which follows where denotes the probability of mortality. A spatial–temporal GAM for modeling the mortality risk of COVID-19 is formulated as follows, where denotes the link function. Three link functions are considered including logistic link function, , that corresponds to the inverse cumulative distribution function (CDF) of the standard logistic distribution, probit link function , where is the standard normal CDF, and complementary log–log (cloglog) function , that is formed from the inverse CDF of the Gumbel distribution. The intercept term is denoted as ; is a row of the design matrix for any strictly parametric model components, such as age groups, gender and history health care use for COVID-19 condition and is the corresponding parameter vector; is modeled as a TPRS function. To simplify the notation, let denote , which takes the form, where and and are coefficients to be estimated, subject to the identifiability constraints . In fitting regression splines, the choice of knot locations and the number of basis functions may have a substantial impact on modeling results. These problems can be avoided by using a relatively large number of basis functions and at the same time imposing a penalty that is designed to ensure that the fitted model is smooth; hence model flexibility is controlled by a smoothing parameter rather than the basis dimension (Ruppert, 2002, Wood, 2000, Wood, 2004, Wood, 2006, Wood, 2008). For the roughness penalty associated with this TPRS basis is, where contains known coefficients and are the parameters. The term denotes the two-dimensional TPRS applied to the geographical locations for the centroids of the neighborhoods where the individuals are from. For ease of presentation, let denote . A two-dimensional TPRS of can be defined as where , , , and are coefficients to be estimated, denotes the Euclidean norm, for and 0 for , subject to identifiability constraints that . The wiggliness penalty functional of is defined as, where contains known coefficients and are the parameters. A major feature of a two-dimensional TPRS is the isotropy of the wiggliness penalty, i.e., smoothing in all directions is treated equally. The isotropy is often considered to be desirable for modeling the interaction between two variables on the same scale, such as geographic coordinates (Wood, 2017). Thin plate regression splines can be seen as a Gaussian process with generalized covariance (Cressie, 1993). Other spatial smoothing techniques could be considered, such as Markov random field (MRF) (Waller et al., 1997, Lawson, 2008, Knorr-Held, 2000). The Bayesian MRF is often fitted using Markov chain Monte Carlo (MCMC) techniques, which may exhibit slow mixing (Christensen et al., 2006). A thin-plate regression spline is applied in this study primarily for its computational efficiency with large datasets (Paciorek, 2007). The interactive effect between neighborhood-level population density and average income, is modeled as a tensor product smoother (Wood, 2006). Tensor product smoother often performs better than isotropic smoother, when the covariates of a smooth are not on the same scale. For ease of presentation, let denote . Tensor product smoother is formulated by firstly representing smooth functions for each of the bivariate variables as and , where and are known basis functions and and are corresponding basis coefficients. In this study, we used the cubic spline basis function. Creating a smooth function of and is achieved by allowing the parameter vary smoothly with , by representing , which gives which allows for neighborhood average income to vary smoothly with population density. The penalty of the interaction term is composed of two components. One applies the penalties of the income to the varying coefficient of the marginal population density smooth , . The second component applies the penalties of the population density smooth to the varying coefficients of the marginal income smooth, , . Then, the roughness of can be measured by the sum of the two penalties where and space are the smooth parameters for and , respectively. The space–time interaction is constructed as a three-dimensional tensor product smooth of space and time (Augustin et al., 2009), which involves firstly specifying the marginal smooth for time and a two-dimensional marginal smooth for space as and , where and are parameters and and are basis functions with , and . The interaction between space and time is constructed by allowing the temporal smooth to varying smoothly within the space dimensions and by specifying which yields where are the coefficients of basis functions of tensor product smooths. The penalty of the space–time interaction term is composed of two components. One applies the penalties of the spatial smooth to the spatially varying coefficient of the marginal temporal smooth , . The second component applies the penalties of the temporal smooth to the temporally varying coefficients of the marginal spatial smooth, , . Then, the roughness of can be measured by the sum of the two penalties where and space are the smooth parameters for space and time, respectively.

Parameter estimation

The model can be written as , where is the link function; is a -vector; is the full design matrix and contains all the parameters need to be estimated. To estimate the parameters , we maximize the penalized log-likelihood conditional on smoothing parameters , operating on the penalty terms , , where denotes the total number of penalty terms. Here is the log-likelihood associated with the Bernoulli response: The smoothing parameters control the tradeoff between the goodness of fit of the model and model smoothness, with , penalizing sharp changes of the splines. To estimate the parameters, penalized likelihood is maximized by a penalized iteratively reweighted least squares (P-IRLS) algorithm. Given the smoothing parameter, at the th P-IRLS iteration, the following penalized sum of squares would be minimized with respect to to find the th estimate : where , and and is proportional to the variance of according to the current estimate . The smoothing parameters are estimated by minimizing the generalized cross-validation score (Craven and Wahba, 1979, Wahba, 1985) for each working penalized linear model of the P-IRLS iteration. The score has the following form: where is the influence matrix (Hastie and Tibshirani, 1990) and is the effective degrees of freedom, or the effective number of parameters, of the model; is a vector of pseudodata and is a diagonal matrix with elements , defined previously, all of these values would be calculated based on estimates of related quantity at each iteration (Hastie and Tibshirani, 1990, Wood, 2004). Defining , based on large sample approximation, the distribution of is derived as, where stands for scale parameter. In logistic regression, . The uncertainty or confidence interval of can be estimated by plugging in the and estimates at convergence of R-IRLS algorithm, along with the smoothing parameters. For more details of estimation for generalized additive models, refer to Hastie and Tibshirani (1990) and Wood (2004).

Model selection and diagnosis

Model selection

Models with different link functions were considered, including logistic, probit, and cloglog link functions. Under each model, model selection was carried out by adding an extra penalty to each smooth term that penalizes functions in the null space of the penalty matrices for each smooth (Marra and Wood, 2011). This is carried out by using the argument select = TRUE in the gam function in R. After model selection, the choice of the link function was evaluated by checking the model fit based on AIC (Akaike, 1973), deviance, percentage of deviance explained, area under the ROC curve (AUC) (Harrell, 2015) and Brier’s score (Rufibach, 2010, Harrell, 2015). AUC measures the discrimination ability of the model. Higher values of the AUC indicate better model discrimination. AUC can examine the ability of the method to distinguish the patients who have the outcome have higher risk predictions than those who do not, but cannot account for calibration, i.e., the magnitude of the disagreement between the observed and predicted responses (Harrell, 2015). To quality how close the predictions are to the actual outcome, Brier’s score (Rufibach, 2010, Harrell, 2015), the mean squared prediction error, is defined as, , where and denote the actual outcome of the event and the predicted probability of the event, respectively. Lower Brier’s scores indicate greater model accuracy. After determining the link function, the relative contribution of each factor for predicting the COVID-19 mortality risk is evaluated by examining the change of model fit by excluding each model term from the full model one at a time.

Residual diagnosis

Residual plots are often used to check the assumptions of regression models; however, residual plots from logistic regression based on deviance or Pearson residuals are generally not informative due to the discreteness of the data (Gelman et al., 2000). Randomized quantile residuals (RQRs) (Dunn and Smyth, 1996) were proposed in the literature to diagnose models for discrete outcomes. However, the power of RQRs for diagnosing discrete models with a limited number of unique values in the outcome variable, such as binary outcome, tends to be low due to the randomness introduced to the estimated cumulative distribution function (Feng et al., 2020). To overcome these challenges, the simulated envelope is recommended to be added to the quantile–quantile (QQ) plot of deviance or Pearson residuals to detect overall departures from the model assumptions as well as outlying observations of the fitted model. For a well-fitted model, most of the residuals are expected to fall within the simulated envelope. The graphical display of residuals versus fitted values or predictors can provide further focused diagnostics. An alternative standard way to make discrete residual plots interpretable is to work with binned residuals (Gelman et al., 2000), which divides the data into categories (bins) of equal size based on their fitted values. Then, the average residual is plotted against the average fitted value for each bin. If the model were true, the residuals should be closer to symmetric about zero, and one would expect about 95% of the residuals to fall inside the error bounds. To examine if there is spatial autocorrelation remained in the residuals, a permutation test for Moran’s I statistic (Moran, 1950, Cliff and Ord, 1981) is used using the spdep package (Bivand and Wong, 2018) in R, calculated by using 999 random permutations of the data over the study region, to establish a distribution of simulated Moran’s I based on a null hypothesis of spatial randomness. The observed value of Moran’s I is then compared with the simulated distribution. The temporal autocorrelation in the residuals was assessed by autocorrelation function (ACF) plot (Box and Jenkins, 1976) and partial autocorrelation (PACF) plots (Venables and Ripley, 2002), which compute autocorrelations for data values at varying time lags. If random, such autocorrelations should be near zero for any time-lags. If non-random, one or more of the autocorrelations will be significantly non-zero.

Predictive accuracy assessment

Using modeling techniques to predict the COVID-19 mortality could help effectively guide public health policy-making to optimize the allocation of the limited resources. In literature, some modeling methods are based on individual-level data (Das et al., 2020, Rosenthal et al., 2020, Xu et al., 2020, Yu et al., 2020) and some are based on aggregated data (Talukder, 2020, Ujiie et al., 2020, Okeahalam et al., 2020, Chaudhry et al., 2020, Karmakar et al., 2021, Lima et al., 2021). The predictive accuracy of the models based on individual versus aggregated data has not been well understood. As a result, this study also examined the accuracy and precision of predictions by comparing the selected best-fitting model based on the individual-level data against commonly used count regression models for modeling the aggregated daily COVID-19 mortality count.

Modeling aggregated daily COVID-19 mortality count

To avoid the potential loss of critical information in the process of data aggregation, the individual-level data were aggregated according to the unique combinations of the levels of the predictors considered in this study, i.e., day, neighborhoods, age groups, gender, ever been hospitalized, ever been in ICU and ever been intubated. The neighborhood-level characteristics, i.e., population density and average income, were linked to this data by neighborhood. By doing so, all the predictors considered in the analysis based on the individual-level data are included in the regression models for modeling the aggregated daily death counts. The aggregation units without COVID-19 positive cases were removed. The final dataset contains 40,882 aggregate units. The outcome variable is the number of deaths within each aggregate unit, which ranges from 0 to 19. The logarithm of the number of positive cases within each aggregate unit is included as an offset term in the count regression models. Modeling count data involves many challenges, especially when the data exhibits a preponderance of zeros and overdispersion. Although alternatives exist for modeling overdispersed count data with an excess of zeros, they are not widely and effectively applied and implemented in standard software. In this study, Poisson, Negative Binomial (NB), and zero-inflated Poisson (ZIP) (Lambert, 1992) models are considered. One major limitation of the Poisson regression is that it assumes the mean and variance conditional on the covariates are the same, which can be restrictive in some applications. NB extends the Poisson by positing that the conditional mean of the response variable is not only determined by the covariates but also a latent component independent of the covariates. Although the NB model could model the overdispersed Poisson data, it is not appropriate for modeling the data with a high percentage of zero counts as in the current context. ZIP is a mixture of two statistical processes, with one always generating “structural” zero counts and the other both “sampling” zero and positive counts. That is, it assumes that each observation comes from one of two potential distributions, with one consisting of a constant zero while the other following Poisson. In a ZIP model, a logit model is typically used to model the probability of the excessive zeros, while the Poisson regression models the rest of the count data.

Evaluation of the prediction performance

The predictive accuracy of the models is evaluated based on both in-sample and out-of-sample prediction in terms of Root Mean Squared Prediction Error (RMSPE) as a measure of the agreement between the observed binary outcome and the predicted probability of the outcome. In-sample analysis means to estimate the model using all available data, and then compare the model’s fitted values to the actual realizations. The out-of-sample predictions are evaluated based on (1) -fold cross-validation, (2) moving window forecasting, (3) leave-one-location-out cross-validation, and (4) leave-one-location-out moving window forecasting cross-validation.

In-sample predictive accuracy check

To ensure the individual and aggregate analyses are comparable, the predicted probabilities of COVID-19 mortality based on the individual-level data analysis for the individuals from the same aggregate unit are summed together, which gives the estimated counts of mortality in that aggregate unit. The RMSPE for the in-sample predictive accuracy is defined as, i.e., where and represent the observed and predicted death counts in the th aggregate unit, respectively, with denoting the total number of aggregate units. For the analysis based on the individual analysis, the observed and predicted death count in the th aggregate unit are and , respectively, where denotes the set of individual observations belonging to the th aggregate unit.

-fold out-of-sample cross-validation

In general, -fold cross validation can be defined as follows: Randomly divide data into equal groups. Let denote the set of data points placed into the th fold. For , train model was fitted on all except th fold. Let denote the resulting predicted probability of the fitted model based on the training dataset , . RMSPE can be then written as

Moving window forecasting predictive check

The predictive ability of the model is further evaluated by its forecasting accuracy, which is evaluated by comparing the forecasts against future realizations of the data. Fig. 2 illustrates the moving window forecast process considered in this study, with , indexing the forecast window, where denotes the number of forecast windows during the study period. denotes the th training dataset, including all observations from the beginning observation period up till the starting time point of the th forecast window. The starting time point of forecasting, , is set as days since the beginning of the study period allowing for enough observations for reliably fitting models to the training dataset. The th moving window forecast dataset is defined as . The forecast horizon, i.e., length of time of forecasting into the future, is set as , and days, respectively.
Fig. 2

An illustration of the moving window forecast process, where and , denote the training and testing datasets, respectively.

The RMSPE at the th forecasting window is defined as where represents the number of observation units in and represent the predicted death count for window based on the train dataset . The average RMSPE over all the forecast windows can be then defined as An illustration of the moving window forecast process, where and , denote the training and testing datasets, respectively.

Leave-one-location-out cross validation

As the data are spatially correlated, leave-one-location-out (LOLO) cross-validation is also used to predict the time series at an unobserved neighborhood. The average RMSPE for LOLO CV across all the neighborhoods is defined as, where denotes the testing data, which is the entire time series of observations at the the neighborhood, , and represents the total number of aggregate units at the th neighborhood. denotes the training dataset for the th neighborhood, which includes the observations from all neighborhoods except for the th neighborhood. Defining each CV fold as the observations taken at one neighborhood ensures that predictions are always being evaluated at a neighborhood excluded from the data on which the prediction was trained.

Leave-one-location-out moving window forecasting cross validation

In Section 3.5.4, the forecasting was made for the entire time series of a location, which does not effectively evaluate the model performance at predicting the future. The RMSPE for evaluating spatial–temporal extrapolation error is therefore examined, which is defined as where denotes the total number of aggregate units at the th neighborhood and the th forecast window; denotes the th forecast window at the th neighborhood and represents the training dataset, which contains the observations prior to the th forecast window after excluding the observations from the th neighborhood, and denotes the predicted disease counts at the th neighborhood and the th forecast window.

Results

After the model selection, all the predictors remain significant regardless of which link function is applied. The results of the model comparison (Table 1) indicate that the model with the probit link function outperforms the models with the logit or cloglog link functions, which yielded the lowest AIC and deviance, highest percentage of deviance explained, highest AUC and lowest Brier’s score.
Table 1

Comparison of the model fits in terms of AIC, deviance (dev), percentage of deviance explained (dev.expl), AUC, and Brier’s score.

LogitProbitCloglog
AIC8175.2738105.8958293.470
dev7942.6987878.7478055.724
dev.expl51.4%51.8%50.7%
AUC0.96680.96720.9661
Brier0.02510.02510.0252
Evaluation of the relative contribution of each predictor for modeling the COVID-19 mortality risk is also conducted by excluding each predictor from the full model. The results (Table 2) indicate that age or history health conditions for COVID-19 are the strongest predictors for mortality status since excluding either of the two variables results in the most substantial change in the measures of model fit. Space, time, and space–time interaction terms also improve the model fit. The model with continuous space–time interaction performs better compared with the model with an additive space–time effect. In addition, accounting for the neighborhood characteristics improves the predictive ability of the model.
Table 2

Evaluation of the contribution of each predictor for modeling the COVID-19 mortality risk by excluding each predictor from the full model, in terms of AIC, deviance, percentage of deviance explained (dev.expl), AUC and Brier’s score.

ModelAICDeviancedev.explAUCBrier’s
Full8105.8957878.74751.8%0.9670.025
-pop × income8145.1497936.67051.4%0.9670.025
-space × time8184.9958087.86150.5%0.9650.026
-gender8191.2747971.68551.2%0.9660.025
-space8252.5058176.20749.9%0.9640.026
-time8352.7768120.09150.3%0.9640.026
-health condition9518.0199355.91642.7%0.9470.028
-age11 115.38410 817.70333.8%0.9250.031
Comparison of the model fits in terms of AIC, deviance (dev), percentage of deviance explained (dev.expl), AUC, and Brier’s score. Evaluation of the contribution of each predictor for modeling the COVID-19 mortality risk by excluding each predictor from the full model, in terms of AIC, deviance, percentage of deviance explained (dev.expl), AUC and Brier’s score. The QQ plots of the deviance residuals for the logistic, probit, and cloglog models with simulated envelopes are displayed in Fig. 3. As displayed, for the logistic and cloglog models, a few residuals at a higher range of the residuals fall outside of the simulated envelopes. By contrast, the points in the QQ-plots based on the probit model mostly fall within the simulated envelopes, which indicates the probit model fits the data well.
Fig. 3

QQ plots of the deviance residuals for the logit, probit and cloglog models. The gray areas are the simulated envelopes based on 1000 simulated data from the fitted model. For a well-fitted model, most of the residuals are expected to fall within the simulated envelope.

The binned residual plots against the fitted values and the predictors are displayed in Fig. 4. The top panels are the average residuals versus average fitted values for each bin under the models with the different link functions. The results revealed that the probit model provided an adequate model fit with almost all the residuals falling within the simulated envelope. By contrast, for logistic and cloglog models, quite a few residuals fall outside of the simulated envelopes. A closer examination of the binned residual plots against the continuous predictors indicates all the models underpredict the mortality risk at one-time point during the second wave, but not substantially. Both the QQ plots and the binned residual plots suggest the probit model provides a better fit to the data as compared to the logistic and cloglog models.
Fig. 4

Binned residual plots of the full models with logit, probit and cloglog link functions, respectively. The gray lines indicate plus and minus 2 standard-error bounds, within which one would expect about 95% of the binned residuals to fall, if the model were true.

QQ plots of the deviance residuals for the logit, probit and cloglog models. The gray areas are the simulated envelopes based on 1000 simulated data from the fitted model. For a well-fitted model, most of the residuals are expected to fall within the simulated envelope. The examination of the temporal and spatial autocorrelation in the residuals of the space–time probit model is displayed in Fig. 5. The ACF and PACF plots do not reveal any significant autocorrelation in the residuals, as shown in the first two panels of Fig. 5. The third panel of Fig. 5 suggests no spatial correlation remained in the residuals with a pseudo -value of 0.953.
Fig. 5

Autocorrelation (ACF) and partial autocorrelation (PACF) plots for examining temporal autocorrelation in residuals and the Monte-Carlo simulation of Moran’s I statistic for examining spatial autocorrelation in residuals (-value 0.953).

Binned residual plots of the full models with logit, probit and cloglog link functions, respectively. The gray lines indicate plus and minus 2 standard-error bounds, within which one would expect about 95% of the binned residuals to fall, if the model were true. Autocorrelation (ACF) and partial autocorrelation (PACF) plots for examining temporal autocorrelation in residuals and the Monte-Carlo simulation of Moran’s I statistic for examining spatial autocorrelation in residuals (-value 0.953).

Estimated model components

The results of the parametric terms for the full model are presented in Table 3 and the non-parametric spline terms for the full model are presented in Fig. 6, Fig. 7. As shown in Table 3, the risk of mortality has no significant difference for age groups below 40–49 years, but the risk increases drastically for the age groups 50–59 years and onward. The mortality risk is significantly higher among males and those who had ever been hospitalized, ever in ICU, or ever intubated.
Table 3

Significance of estimates from the space–time probit model. SE stands for the standard error of the parameter estimate. “edf” represents the effective degrees of freedom of the functional parameters.

Parametric termsEstimateSEp-value
Age (Reference: 19 years and below)
20 to 29 Years−0.4480.4380.306
30 to 39 Years−0.1320.3590.714
40 to 49 Years0.1190.3150.706
50 to 59 Years0.7240.2910.013
60 to 69 Years1.2780.2870.000
70 to 79 Years1.8890.2860.000
80 to 89 Years2.3970.2860.000
90 and older2.8280.2860.000
Gender (Reference: Females)
Male0.3140.0340.000
Others0.1520.1430.288
Health history (Yes vs. No)
Ever Hospitalized0.9380.0370.000
Ever in ICU0.6230.0960.000
Ever Intubated0.5490.1120.000

Smooth termsedfp-value

ft(day)8.184<0.001
fngb(popdensity,income)12.994<0.001
fs(lat,long)25.327<0.001
fst(lat,long,day)53.0690.0028
Fig. 6

The left panel displays the smooth function of day and the right panel shows the smooth function of the population density and average income of the space–time probit model.

Fig. 7

The estimated partial effect of space–time interaction term at specified times on the logit of probability of COVID-19 mortality.

The left panel of Fig. 6 displays the estimated smooth term of the reporting date of COVID-19 on the logit of mortality probability. The plot indicates that the mortality risk increased sharply and peaked around April–May 2020, followed by a decline during summer and then increased again in fall, but the peak is lower than the first peak. The right panel of Fig. 6 displays the two-dimensional tensor product smoother of the population density and average income, which clearly shows their non-linear interactive effect. More specifically, the risk of mortality tends to be much higher for neighborhoods with larger population sizes and lower average incomes. The risk of mortality is the lowest for the neighborhoods with very high average income and moderately low population density. Significance of estimates from the space–time probit model. SE stands for the standard error of the parameter estimate. “edf” represents the effective degrees of freedom of the functional parameters. The spatial–temporal interaction terms (Fig. 7) reflect how the spatial pattern of mortality risk evolves over time after accounting for known risk factors. The results indicate localized differences in the underlying attributes of neighborhoods can have a statistically significant impact on COVID-19 mortality risk. The shifting of the high-risk regions might be attributable to the interventions such as lockdowns and social distancing practices, which reduced the community spread of infection in some areas. Additionally, geographic disparity of health care capability may exist, such as access to clinics, ICU beds, ventilators, or drugs and equipment to effectively treat people most severely affected by COVID-19 infection complications. The left panel displays the smooth function of day and the right panel shows the smooth function of the population density and average income of the space–time probit model. The estimated partial effect of space–time interaction term at specified times on the logit of probability of COVID-19 mortality.

Predictive models based on aggregated data

This section compares the proposed model based on the individual-level data with the count regression models based on the aggregated data constructed based on the individual-level data. Fig. 8 presents the observed versus predicted daily mortality count based on the probit model for modeling the individual-level data and Poisson, NB, and ZIP models for modeling the aggregated data. The results clearly show that the count regression models severely underestimate the higher values of daily mortality count. In other words, count regression models could adequately detect epidemic peaks. In contrast, the probit model provides a much better fit to the data, with the predicted mortality counts much closer to the observed mortality counts. The probit model also under-predicts the higher values of daily mortality counts, but the underestimation is not as pronounced as the count regression models. Fig. 9 presents the QQ-plots of the deviance residuals of the count regression models with simulated envelopes, which confirmed the inadequacy of the count regression models with larger values of residuals falling outside of the simulated envelopes. This phenomenon is more pronounced for Poisson and NB models.
Fig. 8

Predicted versus observed daily mortality counts based on the probit, Poisson, NB and ZIP models, respectively.

Fig. 9

QQ-plots of deviance residuals for Poisson, NB and ZIP models, respectively. The gray areas are the simulated envelopes based on 1000 replicates from the fitted models.

Table 4 presents the parameter estimates of the ZIP model. Similar results were observed based on the Poisson and NB regression, which are not presented here. The results show that all the significant predictors in analysis based on the individual-level data remain significant in the ZIP model, except for the space–time interaction term. As a result, the analysis based on aggregated data cannot reliably elucidate the space–time interactive effect, further demonstrating the potential benefits of the individual-level analysis for gains in power and efficiency for detecting space–time interaction. The inferior performance of the count regression models based on the aggregated data compared to the probit model based on the individual-level data likely stems from the assumptions of distributions for modeling count data.
Table 4

Model estimates of the ZIP model for modeling the daily COVID-19 mortality count. SE stands for the standard error of the parameter estimate. “edf” represents the effective degrees of freedom of the functional parameters.

Parametric termsEstimateSEp-value
Age (Reference: 19 years and below)
20 to 29 Years−0.6521.2930.614
30 to 39 Years0.1291.1200.908
40 to 49 Years1.6770.9550.079
50 to 59 Years2.9930.9220.001
60 to 69 Years4.1370.9180.000
70 to 79 Years5.0840.9170.000
80 to 89 Years5.7880.9160.000
90 and older6.3620.9160.000
Gender (Reference: Females)
Male0.3540.0460.000
Others0.1590.1900.402
Health History (Yes vs. No)
Ever Hospitalized1.1430.0500.000
Ever in ICU0.7310.1120.000
Ever Intubated0.6280.1270.000

Smooth termsedfp-value

ft(day)7.525<0.001
fngb(popdensity,income)7.548<0.001
fs(lat,long)21.434<0.001
fst(lat,long,day)18.8860.1
Predicted versus observed daily mortality counts based on the probit, Poisson, NB and ZIP models, respectively. QQ-plots of deviance residuals for Poisson, NB and ZIP models, respectively. The gray areas are the simulated envelopes based on 1000 replicates from the fitted models. Model estimates of the ZIP model for modeling the daily COVID-19 mortality count. SE stands for the standard error of the parameter estimate. “edf” represents the effective degrees of freedom of the functional parameters.

Evaluation of predictive performance

Table 5 shows that RMSPE based on the in-sample, 10-fold CV, and moving window forecast, leave-one-location out CV and leave-one-location out moving window forecast CV for the probit, Poisson, NB, and ZIP models. The results indicate the probit model had the smallest overall RMSPE as compared to the count regression models. Among the count regression models, the ZIP model had better predictive accuracy than the other methods in most of the cases, except for RMSPE and RMSPE. This implies the performance of the count regression models heavily depends on the schemes of the model validation.
Table 5

RMSPE of the probit model based on the individual-level data and the count regression models (Poisson, NB and ZIP models) based on the aggregated data.

ProbitPoissonNBZIP
RMSPEI0.18150.23810.23880.2336
RMSPECV0.16290.23540.24530.2299
RMSPELOLO0.16970.18160.19020.1841

RMSPEF

7 days0.14220.15070.14910.1511
14 days0.14560.15860.15720.1586
21 days0.15120.16220.16020.1583
28 days0.14590.17680.17520.1728

RMSPEFLOLO

7 days0.11620.12020.11820.1189
14 days0.12570.13420.13240.1341
21 days0.13600.14430.14160.1422
28 days0.13740.15430.15260.1532

RMSPE: RMSPE for in-sample prediction

RMSPE: RMSPE for 10-fold CV

RMSPE: RMSPE for leave-one-location-out CV

RMSPE: RMSPE for moving window forecast

RMSPE: LOLO CV for moving window forecast.

For completeness, we also compared the probit and count regression models for RMSPE and RMSPE over the moving forecasting windows. As displayed in Fig. 10, both RMSPE and RMSPE for all models decrease during summer 2020, followed by an increase during the second wave. The RMSPE and RMSPE for count regression models tend to be higher than the probit model during the second peak time. Such differences increases as the forecasting window size increases. The results imply the count regression models are less accurate for predicting the peak mortality risk as compared to the probit model in this investigation.
Fig. 10

RMSPE (left panels) and RMSPE right panels) for comparing the predictive accuracy between the probit and count regression models for 7 days, 14 days, 21 days and 28 days ahead prediction. Note that in most cases, the RMSPEs based on the count regression models differ only minimally, hence they can hardly be distinguished.

RMSPE of the probit model based on the individual-level data and the count regression models (Poisson, NB and ZIP models) based on the aggregated data. RMSPE: RMSPE for in-sample prediction RMSPE: RMSPE for 10-fold CV RMSPE: RMSPE for leave-one-location-out CV RMSPE: RMSPE for moving window forecast RMSPE: LOLO CV for moving window forecast. RMSPE (left panels) and RMSPE right panels) for comparing the predictive accuracy between the probit and count regression models for 7 days, 14 days, 21 days and 28 days ahead prediction. Note that in most cases, the RMSPEs based on the count regression models differ only minimally, hence they can hardly be distinguished.

Discussion

This study demonstrated a spatial–temporal generalized additive model for modeling and predicting COVID-19 mortality risk in the study region, which could help improve resource allocation across sub-populations. The study adds to the evidence showing the age of 50 years, and onward, males and those who had ever been hospitalized, in ICU, or intubated are at a significantly higher mortality risk due to COVID-19 at the study region during the study period. Neighborhood level population density and average income also have a significant interactive effect on COVID-19 mortality risk. Moreover, we demonstrated a powerful and computationally efficient approach for uncovering the complexity of the inter-correlated nature of space–time COVID-19 data. The identified spatial–temporal effect may be driven by the neighborhood-specific characteristics, such as social network and interactions, regulations and compliances of social-distancing policies, or other socio-demographic information that are not captured by the neighborhood population density and average income. In this study, the model diagnostic indicated the independent error structure was adequate, and no spatial–temporal heterogeneity was identified in the residuals of the best fitting model. However, in some applications, if residuals exhibit spatial or temporal autocorrelation, generalized additive mixed-effect model (GAMM) (Wood, 2006) including random effects with spatial or temporal autocorrelation error structures should be considered. The random effect terms could properly quantify the variability that was unexplained by the predictors in the model. However, the computation for GAMM is typically slower than GAM, and not as numerically robust (Wood, 2006). The space–time smoother is set up using the multi-dimensional tensor product smoother, with the marginal spatial and temporal bases being scale-invariant. This provides flexibility by allowing for different degrees of smoothness for each dimension. The alternative approach of modeling the spatial correlation could also be considered, such as using Markov Random Field (MRF), also called conditional autoregressive prior for spatial random effect term, which is often used for analyzing spatially aggregated data (Waller et al., 1997, Lawson, 2008, Knorr-Held, 2000). Such models are often implemented in the Bayesian framework, which offers the advantage of using prior information to improve inference. In fact, MRF approximation to a thin plate spline that involves only nearby grid cells as neighbors (Rue and Held, 2005, Yue and Speckman, 2010). However, the areal unit problem may arise from using political boundaries that are arbitrarily related to public health. Bayesian models are also more difficult to specify, especially for investigating space–time interaction on a daily basis over a long period. Bayesian estimation methods also require examining prior distributions through a sensitivity analysis. The fitting process may fail in ways that are difficult to diagnose and rectify. The computation can also be intensive or even prohibitive in fitting the model to large datasets. One way to remedy the computational challenge of implementing the Markov Chain Monte Carlo method is to use the Integrated Nested Laplace Approximation (INLA) framework (Rue et al., 2009). INLA uses analytic Laplace approximation and efficient numerical integration schemes to achieve a highly accurate analytical approximation of posterior quantities of interest with relatively short computing times. However, it could be challenging to know when such approximations are adequate. For these reasons, the spatial–temporal generalized additive model by specifying space–time smooth as tensor product smoother offers computational advantage and also significant flexibility for modeling space–time interaction. This study also demonstrates that the individual-level analysis provided a better fit for the data than the analysis based on the aggregated data. The inferior performance of the latter one could be due to the inappropriate distributional assumption of the daily mortality count variable. Generalized additive models for location, scale, and shape (GAMLSS) (Rigby and Stasinopoulos, 2005, Stasinopoulos and Rigby, 2007) are an extension to GAMs that allows all parameters of a certain response distribution to be modeled separately. GAMLSS provides a general framework for fitting regression type models where the distribution of the response variable does not have to belong to the exponential family and includes highly skewed and kurtotic continuous and discrete distribution (Rigby and Stasinopoulos, 2005). Comparing analyses based on individual versus aggregated data based on GAMLSS is interesting to explore as a new project. Further, this study considered three commonly used link functions for binomial regression. Both logit and probit links are symmetric links, which assume that the probability of a given binomial response approaches 0 at the same rate as it approaches 1. The cloglog link function is asymmetric, with the probability of a given binomial response increasing slowly from 0 but then tapering off more quickly as it approaches 1. The results of the model validation and residual diagnosis indicated model with the probit link function provided an adequate fit to the data considered in this study. However, in some other applications, these popular links may not always provide the best fit for a given dataset. Other flexible link function for binomial regression can be chosen, such as a two-parameter class of generalized logistic models (Stukel, 1988, Ghosh and Alzaatreh, 2018), skewed t-link function (Kim et al., 2007), skewed probit link function (Chen et al., 1999). However, these models may not be straightforward to be implemented. Future studies to develop spatial–temporal GAM or GAMLSS with more flexible link functions is therefore warranted. In addition, previous research showed the overall coverage of the GAM model is reasonably close to the normal level, and the component-wise intervals can only be used as a rough guide to the uncertainty of the components, since the intervals are conditional on the smoothing parameters (Wood, 2017, Marra and Wood, 2012). One way to improve the performance of the intervals is to account for the smoothing parameter uncertainty. However, obtaining the distribution of the smoothing parameter is computationally intensive, which requires MCMC simulations or using parameter bootstrap approximation of its sampling distribution (Wood, 2017). Despite the underperformance of component-wise intervals, the performance of all confidence intervals in a GAM setting improves as sample size increases (Wood, 2017, Marra and Wood, 2012). Further, the theoretical argument suggests that component-wise confidence intervals could achieve nominal coverage probability provided that heavy smoothing and collinearity among predictors are avoided. As a result, the empirical coverages of the confidence intervals are expected to be close to the nominal level in this study, since the sample size is fairly large and no heavy smoothing and collinearity among the predictors were identified. However, more rigorous investigations in the context of spatial–temporal GAM through extensive simulation studies would be needed. COVID mortality data may also exhibit a more complex spatial–temporal dependence structure or even discontinuities due to lockdown or distancing policies. Such complexity may not be adequately reflected using tensor product smoothing spline, which may smooth out the abrupt changes. Locally adaptive spatial models proposed in a Bayesian framework (Lee and Mitchell, 2013) could be useful for capturing the spatial discontinuities. Nevertheless, the main constraint of applying the Bayesian model in disease mapping (Lawson, 2008, Knorr-Held, 2000, Lee, 2011) is the computational complexity for large-scale data, which may lead to infeasible computational times. A computationally efficient algorithm for modeling the discontinuity of spatially and temporally correlated infectious disease surveillance data could be developed. Several limitations of this study need to be acknowledged. First, no information on the individual-level socio-economic status, human mobility pattern, and social restrictions at a local level are available, which may contribute to explaining the spatial–temporal pattern identified in the study. Second, survival models to model the time from the date of infection or symptom onset until death or censoring would be a more appropriate approach than logistic regression for modeling mortality risk. However, the time variable provided in the dataset used in this study is a derived/combined variable that best estimates when the disease was acquired and refers to the earliest available date from symptom onset (the first day that COVID-19 symptoms occurred), laboratory specimen collection date, or reported date. Another variable is available in the dataset is called reported date, which is the date on which the case was reported to Toronto Public Health. However, 4513 out of 49,216 (9.2%) of the subjects have episode dates equal to or greater than the reported dates, reflecting the imperfect and incomplete information on the time variables. In addition, death occurrences are subject to reporting delay, and the analysis of deaths by reported date is therefore inevitably distorted by such delay. The lack of accurate information on the dates of symptom onset and mortality thus limits the capacity of carrying out survival analysis in this study.
  5 in total

1.  Effect of socioeconomic factors during the early COVID-19 pandemic: a spatial analysis.

Authors:  Ian W Tang; Verónica M Vieira; Eric Shearer
Journal:  BMC Public Health       Date:  2022-06-18       Impact factor: 4.135

2.  Identifying spatiotemporal patterns of COVID-19 transmissions and the drivers of the patterns in Toronto: a Bayesian hierarchical spatiotemporal modelling.

Authors:  Nushrat Nazia; Jane Law; Zahid Ahmad Butt
Journal:  Sci Rep       Date:  2022-06-07       Impact factor: 4.996

3.  Bayesian disease mapping: Past, present, and future.

Authors:  Ying C MacNab
Journal:  Spat Stat       Date:  2022-01-19

4.  Clustering spatio-temporal series of confirmed COVID-19 deaths in Europe.

Authors:  A Bucci; L Ippoliti; P Valentini; S Fontanella
Journal:  Spat Stat       Date:  2021-10-06

5.  Methods Used in the Spatial and Spatiotemporal Analysis of COVID-19 Epidemiology: A Systematic Review.

Authors:  Nushrat Nazia; Zahid Ahmad Butt; Melanie Lyn Bedard; Wang-Choi Tang; Hibah Sehar; Jane Law
Journal:  Int J Environ Res Public Health       Date:  2022-07-06       Impact factor: 4.614

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.