Literature DB >> 35281183

Interpretable Temporal Attention Network for COVID-19 forecasting.

Binggui Zhou1,2, Guanghua Yang1, Zheng Shi1,2, Shaodan Ma2.   

Abstract

The worldwide outbreak of coronavirus disease 2019 (COVID-19) has triggered an unprecedented global health and economic crisis. Early and accurate forecasts of COVID-19 and evaluation of government interventions are crucial for governments to take appropriate interventions to contain the spread of COVID-19. In this work, we propose the Interpretable Temporal Attention Network (ITANet) for COVID-19 forecasting and inferring the importance of government interventions. The proposed model is with an encoder-decoder architecture and employs long short-term memory (LSTM) for temporal feature extraction and multi-head attention for long-term dependency caption. The model simultaneously takes historical information, a priori known future information, and pseudo future information into consideration, where the pseudo future information is learned with the covariate forecasting network (CFN) and multi-task learning (MTL). In addition, we also propose the degraded teacher forcing (DTF) method to train the model efficiently. Compared with other models, the ITANet is more effective in the forecasting of COVID-19 new confirmed cases. The importance of government interventions against COVID-19 is further inferred by the Temporal Covariate Interpreter (TCI) of the model.
© 2022 Elsevier B.V. All rights reserved.

Entities:  

Keywords:  COVID-19 forecasting; Covariate forecasting; Degraded Teacher Forcing; Multi-task learning; Neural network

Year:  2022        PMID: 35281183      PMCID: PMC8905883          DOI: 10.1016/j.asoc.2022.108691

Source DB:  PubMed          Journal:  Appl Soft Comput        ISSN: 1568-4946            Impact factor:   8.263


Introduction

Worldwide outbreak of coronavirus disease 2019 (COVID-19) has triggered an unprecedented global health and economic crisis. COVID-19 has caused more than 233 million infections and more than 4.7 million deaths worldwide so far.1 Early and accurate forecasts of COVID-19 and evaluation of government interventions are crucial for governments to take appropriate interventions and contain the spread of COVID-19. The forecasting of COVID-19 conducts a prediction of confirmed cases or other indicators caused by COVID-19 in a future horizon. Many works have been proposed to forecast the progression of COVID-19 since its global outbreak. Compartmental models have been widely used for infectious disease modeling ever since its origin in the early 20th century [1]. In [2], Yang et al. employed modified susceptible–exposed–infected–removed (SEIR) model, taking the move-in and move-out parameters into account, to derive the epidemic trend. In [3], Zhou et al. adjusted the SEIR model for predicting epidemic trends by further considering the contact rate and quarantined proportion of COVID-19 transmission. The major limitations of compartmental models are two-fold: 1) compartmental models only consider limited parameters; 2) compartmental models are unable to incorporate covariates for better forecasting. Recently, researchers have proposed several works to eliminate the limitations of compartmental models by combining them with machine learning algorithms [4], [5]. Statistical models, such as Auto-Regressive Integrated Moving Average (ARIMA) and Seasonal Auto-Regressive Integrated Moving Average (SARIMA), were applied to forecast the epidemiological trends of the COVID-19 pandemic in [6], [7], [8]. To improve forecasting accuracy, exogenous variables were considered in recent studies for COVID-19 forecasting [9], [10]. Statistical models normally rely on multiple assumptions, which limits their applications. In addition to compartmental models and statistical models, machine learning models and deep learning models have also been widely used in the COVID-19 forecasting problem [11], [12], [13]. In [3], the authors proposed a logistic growth model for near-term predictions. In [14], Ardabili et al. investigated several machine learning models, including the multi-layered perceptron and the adaptive network-based fuzzy inference system, to predict the outbreak of COVID-19. As for deep learning models, long short-term memory (LSTM) networks [2], [15], [16], [17], gated recurrent unit (GRU) networks [16], convolutional neural networks (CNN) [15], [17], [18], attention-based networks (e.g., Transformer [17], [19], [20], dot-product attention models [21], graph attention networks [22]), etc., were applied to forecast the COVID-19. In addition, hybrid approaches that combine different deep learning methods, such as CNN-LSTM [23], were also investigated for better forecasting of COVID-19. To the best of our knowledge, although machine learning models and deep learning models show better model capacity and flexibility in learning from covariates and forecasting epidemic progression, they suffer from unsatisfying performance of long-term forecasting, overfitting, and poor interpretability. To overcome the aforementioned limitations, we propose the Interpretable Temporal Attention Network (ITANet) for interpretable multivariate time series forecasting and apply it to the forecasting of COVID-19 epidemic. The model is with an encoder–decoder architecture, where both the encoder and the decoder employ LSTM for extracting temporal features and multi-head attention for capturing long-range dependencies. The model simultaneously takes historical information, a priori known future information, and pseudo future information into consideration. The pseudo future information is learned with the covariate forecasting network (CFN) and multi-task learning (MTL). In addition, We further propose the degraded teacher forcing (DTF) method to train the model efficiently. The main contributions of this work can be summarized as follows: The proposed ITANet has a superior model architecture and learning capacity for time series forecasting and covariate importance extraction; The proposed multi-task learning paradigm has the ability to provide additional supervision information for the model to achieve better forecasting performance. The proposed degraded teacher forcing method is capable of training the model efficiently and mitigating train-test mismatch. The proposed model is capable of providing promising performance in forecasting of the COVID-19, and evaluating the importance of government interventions, which is beneficial for governments to contain the progression of COVID-19. The remainder of this paper is organized as follows. In Section 2, we introduce the related works of this study. In Section 3, we introduce the architecture and main components of ITANet, as well as the degraded teacher forcing method and the multi-task learning paradigm. In Section 4, we introduce experimental datasets, data preprocessing and experimental settings for comparing our model with existing deep learning models in the forecasting of COVID-19. In Section 5, we show the experimental results of forecasting performance, model uncertainty, model complexity, and ablation studies. We further conduct extensive experiments to demonstrate the benefits of DTF and MTL on the forecasting performance and the benefits of DTF on other models with encoder–decoder architectures. We also evaluate the importance of government interventions with the help of temporal covariate interpreter. Finally, we provide a detailed discussion and conclude this work in Section 6 and Section 7, respectively.

Related works

COVID-19 forecasting

Multiple types of forecasting models have been applied to forecasting the progression of COVID-19, including compartmental models [2], [3], statistical models [6], [7], [8], and deep learning models [16], [18], [20]. However, compartmental models and statistical models suffer from limited expression capability and unsatisfying forecasting performance, while deep learning models are usually blamed for overfitting issues and poor interpretability.

Incorporating and interpreting covariates

To improve forecasting accuracy, researchers have made huge efforts to incorporate covariates so that more information could be utilized. Even for compartmental models and statistical models, recent studies have proposed to incorporate covariates (e.g., changes in the policies [5], number of currently hospitalized patients [10]) to enhance the models’ performance. For deep learning models, it would be easier to incorporate covariates (e.g., mobility information [20]). However, these works usually simply combined all incorporated covariates together and inputted them to the networks. Despite some recent studies [24], how to make full use of the covariates and how to interpret the importance of incorporated covariates remain further investigation.

Encoder–decoder models and their training strategies

Encoder–decoder models have been widely employed for time series forecasting in recent years [25], [26]. Encoder–decoder models are usually difficult to train, especially when encoder time steps or decoder time steps are long. A well applied training strategy for encoder–decoder models is Teacher Forcing [27], where the decoder has access to all the observed target values at every decoding step to mitigate error propagation. However, since the model cannot always access all the observed target values during the inference stage in a multi-horizon forecasting problem, there inevitably exists a gap between the input information for training and inference, which is the so called train-test mismatch. More efficient training strategies should be proposed to train encoder–decoder models well while mitigating train-test mismatch.

Proposed method

Interpretable temporal attention network

As shown in Fig. 1, the proposed ITANet is with an encoder–decoder architecture consisting of an encoder for processing historical information and a decoder for processing future information and forecasting. From the very beginning of the encoder and decoder, the input transformation layer transforms the inputs to high-dimensional features, and the temporal covariate interpreter (TCI) layer interprets the importance of each input covariate, respectively. After that, an LSTM layer for extracting temporal features and a multi-head self-attention layer for capturing long-range dependencies are connected to the TCI layer on both the encoder side and decoder side. In addition, the decoder also includes three components: a covariate forecasting network (CFN) for expanding the future inputs, an encoder–decoder attention layer for attending the encoder representation, and a linear layer for final linear regression.
Fig. 1

The network architecture of ITANet.

The network architecture of ITANet. The model takes three types of covariates as inputs, i.e., covariates with historical information only (), covariates with historical information and pseudo future information () and covariates with historical information and a priori known future information (). The pseudo future information are generated by the covariate forecasting network. Assuming a -horizon forecasting problem, the original input lengths of , and are , and , respectively.

Covariate forecasting network

Incorporating more useful covariates and their future information into time series forecasting helps to improve the forecasting performance of the model. Traditional encoder–decoder architectures usually take only the value of the target sequence at the last time step before the start point of forecasting, i.e., , as the input on the decoder side. In the proposed encoder–decoder architecture, we have incorporated covariates with a priori known future information into the inputs on the decoder side to achieve potential forecasting performance improvement. However, covariates that we have a priori on are limited. Therefore, in this section, we posit that by introducing some critical and predictable covariates to the inputs on the decoder side, the performance of the model can also be promoted. To that end, we propose the covariate forecasting network to generate pseudo future information for some covariates with their historical information and expand the inputs on the decoder side: To avoid overfitting to the historical data, the covariate forecasting network is a lightweight encoder–decoder architecture composed of a single LSTM layer on both the encoder and decoder sides, instead of a complex network architecture.

Input transformation layer

The covariates considered are either categorical variables or numerical variables. For categorical variables, we successively apply label encoding and linear transformation, to transform each categorical variable to a -d representation vector. Here the dimension is determined by an empirical rule [28]: where and are the number of categories and the predefined maximal embedding size for the th categorical variable, and the function returns a rounded integer number. Through trial and error, the 1.6 and 0.56 in Eq. (2) empirically give good embedding sizes for categorical variables. For numerical variables, we simply apply linear transformation to map the original value to a -d representation vector. After obtaining the representations for each categorical variable and numerical variable, an additional linear transformation is applied to transform each representation vector into a -d hidden representation space.

Temporal covariate interpreter

Inspired by the variable selection network in [24], we build the temporal covariate interpreter to interpret the importance of the input covariates. Considering the complexity and expression capability of the network, the temporal covariate interpreter is composed of a 2-layer feed forward network and a activation. Let denote the transformed representation of the th covariate, covariate-wise importance vector at time step , i.e., , is generated as: and where is the Exponential Linear Unit activation function [29], , , and , are learnable weights and biases of the first and second linear layer, respectively. is the total number of covariates, and the operation is to generate a flatten vector across all covariates. After obtaining the covariate-wise importance vector, the output of the TCI layer at time step can be expressed as: where is the th element of .

LSTM layer

The long short-term memory (LSTM) architecture was originally designed to process long sequential data and remember information for a longer period than the original recurrent neural network (RNN) [30], [31]. An LSTM unit has three gates, i.e., forget gate, input gate, and output gate, that control the propagation of two kinds of information, i.e., the cell state and the hidden state. Here we use an LSTM layer to process the representations of inputs to extract temporal features, where the hidden states of all time steps from the LSTM layer are passed to the latter multi-head self-attention layer. The hidden states also provide additional positional information for the attention modules, so that no additional positional encoding is required as inputs to the encoder and decoder.

Multi-head attention layers

Although the LSTM layer is able to extract temporal features and provide positional information, the long-range dependencies between the inputs are still difficult to be captured. To solve the problem, attention mechanism is introduced. Let , , denote the fundamental constituents, i.e., the Query, Key and Value matrices, for calculating the scaled dot-product attention [32], which is specified as: where is the input dimension. Multi-head attention, as an extension of the scaled dot-product attention, is more effective and is employed in our model: and where , , , as well as , are learnable weights which conduct linear transformations from the original representation space to different sub-spaces, and is the number of attention heads. Multi-head attention is expected to learn from those representation sub-spaces and obtain various features, and thus enhance the model performance. Following the multi-head attention module, there is also a feed forward network with two linear transformations and one ELU activation in between. Also, a residual connection around each of the multi-head attention module and the feed forward network is employed. On the encoder side, a multi-head self-attention layer is adopted, where the hidden states from the encoder-side LSTM layer form the , , matrices. On the decoder side, a multi-head self-attention layer and a encoder–decoder attention layer are adopted. Similarly, the hidden states from the decoder-side LSTM layer form the , , matrices to the decoder-side multi-head self-attention layer. While for the encoder–decoder attention layer, the , matrices come from the outputs of the encoder-side multi-head self-attention layer and the matrix comes from the outputs of the decoder-side multi-head self-attention layer. It is worth mentioning that to avoid the information leaking from the future steps, masked multi-head attention is employed on the decoder side, which includes an extra look-ahead mask compared to the module on the encoder side.

Encoder–decoder architecture

Encoder–decoder architectures have been employed in processing sequential data since the era of recurrent neural networks. Recently, as the development of attention mechanism, the application of encoder–decoder architectures is becoming even broader. For multi-horizon forecasting, the decoder, either of an RNN model or of an attention-based model, normally conducts one-step training and step-by-step inference. During the inference phase, the decoder can only access the last target value before the start point of decoding, therefore the latter decoding steps would be conditioned on the former predictions of the decoder itself. Therefore, the step-by-step inference is not quite effective, especially when the forecasting horizon is long. In this paper, we employ the generative inference proposed in [33] for one-step inference: the decoder is fed with the past -step target values, and outputs the -step forecasting results with one decoding step. In summary, the forecasting results are generated as follows: and where the future part of , i.e., , is given by the covariate forecasting network as Eq. (1) shows.

Degraded teacher forcing

Traditionally, models with encoder–decoder architectures can be trained with Teacher Forcing [27], in which the decoder has access to all the observed target values at every decoding step to mitigate error propagation. However, since the model cannot always access all the observed target values during the inference stage in a multi-horizon forecasting problem, there inevitably exists a gap between the input information for training and inference. In other words, teacher forcing has advantages in properly training the model, but also leads to train-test mismatch. To mitigate train-test mismatch, we propose the degraded teacher forcing for better training the proposed model. Instead of training the model with one-step shifted target sequences as the teacher forcing does, the model is trained with the degraded forcing sequences (DFS). Applying DTF requires 3 types of forcing sequences, i.e., 0-padding sequences, -step shifted target sequences and one-step shifted target sequences. The Type 1 DFS (0-padding sequences) are to mimic the situations that no target values could be observed. The Type 2 DFS (-step shifted target sequences) are to mimic the information to the ITANet in the testing stage. The Type 3 DFS (one-step shifted target sequences) are used to force the training so that the model can be properly trained. The combination of the three types of Degraded Forcing Sequences, i.e., the ratios to use the three types of DFS, is further determined by hyper-parameter tuning. When DTF is applied, the training samples should be generated at the very beginning of each iteration by the sliding window method, as presented in Alg. 1. For generating validation and testing samples, we simply adopt Type 2 DFS for all samples for one-step inference as we introduced before.

Multi-task learning

As illustrated before, the covariate forecasting network is introduced to generate pseudo future information for some useful covariates. To guarantee the quality of the generated pseudo future information, and to train the forecasting model properly, multi-task learning is introduced to train the whole ITANet with additional supervision from pseudo future known covariates, comparing to the original single-task supervised learning. Let the task of forecasting COVID-19 confirmed cases be the primary task, and the tasks of forecasting pseudo future known covariates be the auxiliary tasks. The primary task is optimized by minimizing the sum of quantile losses [34] w.r.t multiple quantiles and all time points to forecast: , where The quantile loss is to offer precise numerical forecasting results and insights of forecast uncertainties by providing multiple forecasting intervals. The commonly used quantiles include: 0.02, 0.1, 0.25, 0.5, 0.75, 0.9, and 0.98. The auxiliary tasks are optimized by common regression losses (e.g., mean squared error loss, mean absolute error loss, etc.) or common classification losses (e.g., binary cross entropy, categorical cross entropy, etc.) accordingly. For each auxiliary task across all time points to forecast, the loss function is determined by: The whole ITANet, denoted as , is then jointly optimized as follows: , where is the number of auxiliary tasks, is the weight for balancing auxiliary tasks and the primary task, and is the joint loss function. To sum up, the procedures for training the proposed ITANet with degraded teacher forcing and multi-task learning is shown in Alg. 2. Note that the training set is generated at each iteration with variations, while the validation set is generated beforehand and keep unchanged to accelerate the validation process.

Experiments

Experimental datasets

The variables-of-interest in the experiments, including input variables (i.e., government interventions, and air quality related variables) and target variables (i.e., new confirmed cases), were picked up from the following datasets: Confirmed Cases, Confirmed Deaths, and Government Interventions: Oxford COVID-19 Government Response Tracker [35] is a tool to rigorously and consistently track policy responses around the world. Government interventions, including pharmaceutical interventions and non-pharmaceutical interventions, are divided into four categories: containment and closure policies, economic policies, health system policies and miscellaneous policies. The dataset provides daily updated government interventions, confirmed cases and deaths from more than 180 countries and regions. The dataset also provides state-level data for some countries including the United States. Note that miscellaneous policies are excluded since they are basically described with free text. Temperature and Air Quality Index: The temperature and air quality index data from the World Air Quality Index project. The dataset provides the values for each air pollutant species (e.g, PM10, PM2.5) and meteorological data (e.g., wind, precipitation, temperature), covering about 380 major cities in the world. Specifically, the exact variables taken in the experiments for three types of inputs (i.e.,  (variables with historical information only), (variables with historical information and pseudo future information), (variables with historical information and a priori known future information)) are shown in Table 1:
Table 1

The exact variables taken in the experiments for three types of inputs.

xhHistorical new confirmed cases, 19 government interventions

xpfDate index, month index, state code

xrfTemperature, air quality index
The exact variables taken in the experiments for three types of inputs. Hyper-parameters, corresponding tuning spaces, and best hyper-parameter settings of the ITANet for the three states. : mapped vector dimension of numerical variables; : dimension of hidden representations (including combined hidden representation, hidden states of LSTM layers in the CFN or the main encoder–decoder network) and feed-forward size (inner dimension of feed forward networks within the TCI or following multi-head attention modules); : number of attention heads; : dropout rate; (, , ): ratios to use 3 types of DFS.

Data preprocessing

The data extracted from above datasets cannot be directly used for model training or analysis. Missing values and abnormal values are common in raw data, and thus further data filling and cleaning is required. For numerical variables, we filled the blanks with 0. For categorical variables, we filled the blanks with forward filling method, which forwarded the last observed value to the blanks till the next observed value, considering that these categorical variables may last for a few days. We checked the codebook of the government intervention dataset to make sure there was no abnormal values in these variables. On considering the data availability, we conducted the experiments on 3 states of the United States: Illinois, California and Texas. The time series between April 1st, 2020 and April 28th, 2021 was chosen and split into three parts: training set, from April 1st, 2020 to March 31st, 2021; validation set, from April 1st, 2021 to April 14th, 2021; testing set, from April 15th, 2021 to April 28th, 2021. Sliding window method was applied to generate training, validation and testing samples as Alg. 1 shows. Since the available data points (in days) were limited, we slid the window by a stride=1. To avoid the training being dominated by some large scale numerical variables, pre-scaling was conducted on all numerical variables. The pre-scaler for each variable was generated only with the data in the training set. The pre-scaling was done by Z-score normalization, where population mean and population standard deviation were firstly calculated, and then the original data is transformed according to the following equation:

Experimental settings

We train the model for 14-horizon forecasting with 28-horizon inputs under the multi-task learning paradigm and the joint loss was calculated as Eq. (14). Here we used quantiles including: 0.1, 0.5, and 0.9. Due to data availability, the auxiliary tasks include forecasting the future temperature and Air Quality Index (AQI). We set for all the auxiliary tasks. Mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE) were introduced as the performance evaluation metrics: , where and denote the true value and predicted value of new conformed cases at time t, and is the length of time horizons to forecast. We compared the proposed ITANet with the other four deep learning models, including CNN [18], LSTM [16], Transformer [20] and Temporal Fusion Transformer (TFT) [24]. As naive compartmental methods and statistical methods cannot incorporate covariates, they were not considered in comparison. The models were implemented, trained and tested in Python with TensorFlow 1.15. The models were trained over 100 epochs and the batch size was set to 256. The models were trained and tested on a HPC with 2 Intel(R) Xeon(R) Gold 6230 CPUs (@2.10 GHz), 256 GB RAM, and 2 NVIDIA V100 32 GB GPUs. Grid search was employed to determine the hyper-parameters of the neural networks. The hyper-parameter tuning spaces for the proposed model are listed in Table 2. The model with minimum validation loss was selected as the best model to conduct the following evaluations and comparisons on the testing set.
Table 2

Hyper-parameters, corresponding tuning spaces, and best hyper-parameter settings of the ITANet for the three states.

Hyper-parameterTuning SpaceBest for CABest for ILBest for TX
dn4, 8444
h4, 8, 16, 32, 64163216
H1, 2, 4, 8, 16, 32, 6481616
pdrop0.1, 0.2, 0.3, 0.4, 0.50.50.50.3
(p1, p2, p3)(0.1, 0.1, 0.8), (0.15, 0.15, 0.7), (0.25, 0.25, 0.5)(0.15, 0.15, 0.7)(0.15, 0.15, 0.7)(0.25, 0.25, 0.5)

: mapped vector dimension of numerical variables;

: dimension of hidden representations (including combined hidden representation, hidden states of LSTM layers in the CFN or the main encoder–decoder network) and feed-forward size (inner dimension of feed forward networks within the TCI or following multi-head attention modules);

: number of attention heads;

: dropout rate;

(, , ): ratios to use 3 types of DFS.

Experimental results

Performance evaluation

We compared the proposed ITANet with the other four baseline models, i.e., CNN, LSTM, Transformer and TFT, evaluated on the forecasting of the COVID-19 new confirmed cases of 3 states of the United States, i.e., California, Illinois and Texas. Experimental results are shown in Table 3 and Fig. 2.
Table 3

Forecasting performance of ITANet and baseline models.

ModelStateMAERMSEMAPE
CNNCalifornia1068.151273.530.5808
Illinois790.64900.450.3233
Texas2704.022932.471.1087

LSTMCalifornia3986.684170.142.5184
Illinois809.13924.780.3135
Texas1179.961584.900.6166

TransformerCalifornia658.54891.260.5413
Illinois848.7983.080.2862
Texas1595.422049.380.4353

TFTCalifornia993.011191.830.7656
Illinois694.77840.820.2274
Texas1282.841664.100.6281

ITANetCalifornia597.69697.440.3950
Illinois294.46392.490.1215
Texas875.441073.590.3462
Fig. 2

Examples of COVID-19 forecasting results from Apr. 15 to Apr. 28 by ITANet and other models.

The ITANet outperforms all the other models in terms of all evaluation metrics, i.e., MAE, RMSE and MAPE. From Fig. 2, we can find that the proposed model is better than the others not only in numerical regression, but also in trend prediction. Forecasting performance of ITANet and baseline models. Examples of COVID-19 forecasting results from Apr. 15 to Apr. 28 by ITANet and other models.

Model uncertainty

In order to study model uncertainty, we trained the proposed ITANet with quantile loss as defined in Eq. (11), so that the models provided prediction intervals in addition to precise value predictions. Since the model was trained with 0.1, 0.5, 0.9 quantiles, we use 0.1 quantile prediction and 0.9 quantile prediction as the lower bound and upper bound of the prediction intervals, respectively. The prediction intervals of new confirmed cases for 14-horizon forecasting are shown in Fig. 3.
Fig. 3

Prediction intervals for COVID-19 forecasting by ITANet in California, Illinois and Texas.

From Fig. 3, we can find that the prediction intervals given by our model can accurately reflect the maximal and minimal scales of the progression of the COVID-19 epidemic. The prediction intervals can help governments to be fully prepared for the epidemic, which is of great significance. Prediction intervals for COVID-19 forecasting by ITANet in California, Illinois and Texas. Model complexity of ITANet and baseline models.

Model complexity

To compare the complexity of the proposed model with the baseline models, the total number of parameters and floating-point operations (FLOPs) of each model are listed in Table 4. It can be seen from Table 4 that with the well designed network architecture, the proposed ITANet is very efficient in exploiting input information and achieves the best performance with the least computational complexity.
Table 4

Model complexity of ITANet and baseline models.

ModelStateNo. of ParamsFLOPs
CNNCalifornia146.010k291.476k
Illinois90.936k181.332k
Texas149.482k298.228k

LSTMCalifornia53.488k106.986k
Illinois53.664k107.338k
Texas390.666k189.184k

TransformerCalifornia48.368k97.583k
Illinois48.544k97.935k
Texas48.384k97.675k

TFTCalifornia58.896k114.970k
Illinois123.062k274.924k
Texas786.432k1610.574k

ITANetCalifornia32.230k70.652k
Illinois32.406k71.004k
Texas32.438k71.212k

Ablation studies

In this work, we proposed a covariate forecasting network with multi-task learning (MTL) to provide additional supervision for training a model with better forecasting performance. In addition, we also proposed the degraded teacher forcing (DTF) method to train the model efficiently while reducing the train-test mismatch. In this section, we conducted ablation studies to investigate the advantages of multi-task learning and degraded teacher forcing. For comparison, three models, i.e., ITANet without MTL, ITANet without DTF, and ITANet without MTL & DTF, were evaluated. Table 5 shows the performance of the three models in forecasting COVID-19 new confirmed cases in California, Illinois and Texas under the aforementioned configurations.
Table 5

Ablation studies for multi-task learning and degraded teacher forcing.

ModelStateMAERMSEMAPE
ITANetCalifornia597.69697.440.3950
Illinois294.46392.490.1215
Texas875.441073.590.3462

ITANet w/o MTLCalifornia1097.591351.520.6971
Illinois521.14598.910.1758
Texas1313.541621.780.4554

ITANet w/o DTFCalifornia944.751101.760.5077
Illinois863.37985.630.3573
Texas1206.081367.080.4676

ITANet w/o MTL&DTFCalifornia1663.991903.210.7683
Illinois858.09957.540.3497
Texas1913.672230.830.5367
From Table 5, we can find that both MTL and DTF have benefits to the COVID-19 forecasting. Without MTL or DTF, the performance of the model drops significantly with a few exceptions. For training without DTF, the performance loss may due to that the forecasting horizons are too long. The accumulated training errors make the model difficult to train. When DTF is applied, the model is trained with the guidance of the degraded forcing sequences to mitigate the training errors. At the same time, due to our elaborate design of the degraded forcing sequences and the training procedure, the DTF also avoids great train-test mismatch while training the model efficiently. As for the studies for MTL, we posited that to provide highly relevant future information may help the forecasting, so the performance loss may due to the reduction of supervision information or network parameters. Ablation studies for multi-task learning and degraded teacher forcing. From Table 5, we can also find that DTF is more useful than MTL. The possible reason was that we only provided future temperature and AQI information for supervision considering data availability, which may have limited contributions to the outputs. However, the key idea behind this method provides a direction for the use of future information for additional supervision by combining the covariate forecasting network and multi-task learning. Given sufficient useful covariates, the method is expected to contribute more. The importance of government interventions against COVID-19 in California, Illinois and Texas . Since DTF is also suitable for other models with encoder–decoder architectures, such as TFT, LSTM and Transformer, we also compared the differences in forecasting performance between models trained with and without DTF. The results in Table 6 demonstrate that the proposed DTF training method is effective and can achieve a consistent performance improvement on the models with encoder–decoder architectures.
Table 6

Effectiveness of degraded teacher forcing for other models with encoder–decoder architectures.

ModelStateMAERMSEMAPE
LSTMCalifornia3986.684170.142.5184
Illinois809.13924.780.3135
Texas1179.961584.900.6166

LSTM w/ DTFCalifornia886.511128.620.6780
Illinois796.26911.110.3202
Texas917.861304.030.3110

TransformerCalifornia658.54891.260.5413
Illinois848.70983.080.2862
Texas1595.422049.380.4353

Transformer w/ DTFCalifornia677.26821.800.5342
Illinois460.62552.170.1585
Texas921.791133.110.4028

TFTCalifornia993.011191.830.7656
Illinois694.77840.820.2274
Texas1282.841664.100.6281

TFT w/ DTFCalifornia1426.351653.070.7120
Illinois525.92645.010.2209
Texas1342.251563.980.5842
Effectiveness of degraded teacher forcing for other models with encoder–decoder architectures. The importance ranking scores of government interventions.

Covariate importance

One of the motivations of this work is to interpret covariate importance in time series forecasting. As for the forecasting of COVID-19, we hope to determine which interventions are most important and thus provide a guidance for future epidemic containment. We consider 3 categories of government interventions, i.e., containment and closure policies, health system policies, and economic policies, in our experiments, which are denoted by ’Ca_b’, ’Ha_b’, ’Ea_b’, where ’a’ is the order of a policy within each category of interventions, and ‘b’ is the name of the policy. Fig. 4 shows the importance of each government intervention regarding the three states obtained by the temporal covariate interpreter, where the importance is measured by importance in percentage (i.e., the sum of the importance of all covariates is 100 percentages). Note that the importance of those non-policy covariates is not shown in Fig. 4.
Fig. 4

The importance of government interventions against COVID-19 in California, Illinois and Texas .

To better evaluate the importance of government interventions, we scored three categories of interventions, i.e., containment and closure policies, economic policies, health system policies, according to their importance within each category, and summed the scores obtained in three states as the final importance score of each intervention. The scores of all the government interventions are shown in Table 7.
Table 7

The importance ranking scores of government interventions.

Intervention CategoryInterventionsCaliforniaIllinoisTexasTotal Score
Containment and closure policiesC1_School closing82414
C2_Workplace closing14611
C3_Cancel public events61815
C4_Restrictions on gatherings28212
C5_Close public transport37515
C6_Stay at home requirements73313
C7_Restrictions on internal movement46111
C8_International travel controls55717

Health system policiesH1_Public information campaigns3115
H2_Testing policy42410
H3_Contact tracing73515
H4_Emergency investment in healthcare56213
H5_Investment in vaccines27716
H6_Facial Coverings14611
H7_Vaccination Policy65314

Economic policiesE1_Income support3429
E2_Debt/contract relief1135
E3_Fiscal measures2248
E4_International support4318
From Table 7, we can find that international travel controls, canceling public events and closing public transport are among the top 3 important containment and closure policies. Investment in vaccines, vaccination policy and contact tracing are considered as the most important health system policies. For economic policies, the scores are not consistent for different states, but may imply that income support and debt/contract relief might be the most and least important measures, respectively. It is worth mentioning that the importance of a government intervention can be affected by a lot of factors, such as when the intervention starts, how stringent it is, how long it lasts, etc, resulting in the differences of the importance of government interventions between the states. We expect that these findings can provide valuable suggestions for the governments to choose more effective control measures to contain the progression of the COVID-19.

Discussion

In this work, we propose a novel forecasting model, i.e., ITANet, as well as a set of methods to achieve better COVID-19 forecasting performance. According to the experimental results, the proposed ITANet outperforms the other baseline models in 14-day COVID-19 forecasting for three US states. More importantly, the proposed model is able to reveal the importance of covariates including government interventions through the temporal covariate interpreter. This capability provides an evaluation method on the government interventions in containing the progression of COVID-19. The proposed model makes full use of various kinds of information, including historical information, a priori known future information, and pseudo future information learned with the covariate forecasting network and multi-task learning. The proposed degraded teacher forcing method is able to train the model efficiently. It is worth mentioning that the proposed model is applicable for other time series forecasting problems, in addition to the forecasting of COVID-19 confirmed cases, which may provide satisfying performance of long-term forecasting and interpretable covariate importance. The DTF method is also applicable for training other models with encoder–decoder architectures. One limitation of this work is that although the proposed ITANet achieves the best performance in forecasting COVID-19 confirmed cases, it is not able to always keep satisfying forecasting performance at each time step within the 14-day forecasting horizons. This may be due to the lack of training data or the use of sliding window method to cut the original time series, resulting in insufficient feature capture. The model can be further improved to support longer-term COVID-19 forecasting by designing better attention mechanisms. Another limitation is that although the proposed model is able to interpret the covariate importance at each forecasting horizon, the actual effectiveness of a government intervention is not provided. It is complicate to determine the actual effectiveness of a government intervention, since it can be affected by a lot of factors (e.g., duration, strictness of the intervention). It is our future direction to fully determine the effectiveness of these interventions.

Conclusion

In this work, we proposed a method based on ITANet, a novel deep neural network, to conduct COVID-19 forecasting and infer the importance of government interventions. We used the covariate forecasting network and multi-task learning paradigm to introduce more supervision information for training the model to improve forecasting performance. We further proposed the degraded teacher forcing method to train the model efficiently while mitigating train-test mismatch. The ITANet was compared with other deep learning models, including CNN, LSTM, Transformer and Temporal Fusion Transformer. The experimental results demonstrated the effectiveness of our proposed model in the forecasting of COVID-19.

CRediT authorship contribution statement

Binggui Zhou: Data curation, Methodology, Software, Formal analysis, Writing – original draft. Guanghua Yang: Conceptualization, Supervision, Writing – review & editing. Zheng Shi: Validation, Writing – review & editing. Shaodan Ma: Supervision, Writing – review & editing.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  13 in total

1.  Learning to forget: continual prediction with LSTM.

Authors:  F A Gers; J Schmidhuber; F Cummins
Journal:  Neural Comput       Date:  2000-10       Impact factor: 2.026

2.  A global panel database of pandemic policies (Oxford COVID-19 Government Response Tracker).

Authors:  Thomas Hale; Noam Angrist; Rafael Goldszmidt; Beatriz Kira; Anna Petherick; Toby Phillips; Samuel Webster; Emily Cameron-Blake; Laura Hallas; Saptarshi Majumdar; Helen Tatlow
Journal:  Nat Hum Behav       Date:  2021-03-08

3.  Prediction of COVID-19 Confirmed Cases Combining Deep Learning Methods and Bayesian Optimization.

Authors:  Hossein Abbasimehr; Reza Paki
Journal:  Chaos Solitons Fractals       Date:  2020-11-28       Impact factor: 5.944

4.  Forecasting the dynamics of cumulative COVID-19 cases (confirmed, recovered and deaths) for top-16 countries using statistical machine learning models: Auto-Regressive Integrated Moving Average (ARIMA) and Seasonal Auto-Regressive Integrated Moving Average (SARIMA).

Authors:  K E ArunKumar; Dinesh V Kalaga; Ch Mohan Sai Kumar; Govinda Chilkoor; Masahiro Kawaji; Timothy M Brenza
Journal:  Appl Soft Comput       Date:  2021-02-08       Impact factor: 6.725

5.  Forecasting COVID-19 cases: A comparative analysis between recurrent and convolutional neural networks.

Authors:  Khondoker Nazmoon Nabi; Md Toki Tahmid; Abdur Rafi; Muhammad Ehsanul Kader; Md Asif Haider
Journal:  Results Phys       Date:  2021-04-19       Impact factor: 4.476

6.  STAN: spatio-temporal attention network for pandemic prediction using real-world evidence.

Authors:  Junyi Gao; Rakshith Sharma; Cheng Qian; Lucas M Glass; Jeffrey Spaeder; Justin Romberg; Jimeng Sun; Cao Xiao
Journal:  J Am Med Inform Assoc       Date:  2021-03-18       Impact factor: 4.497

7.  Modified SEIR and AI prediction of the epidemics trend of COVID-19 in China under public health interventions.

Authors:  Zifeng Yang; Zhiqi Zeng; Ke Wang; Sook-San Wong; Wenhua Liang; Mark Zanin; Peng Liu; Xudong Cao; Zhongqiang Gao; Zhitong Mai; Jingyi Liang; Xiaoqing Liu; Shiyue Li; Yimin Li; Feng Ye; Weijie Guan; Yifan Yang; Fei Li; Shengmei Luo; Yuqi Xie; Bin Liu; Zhoulang Wang; Shaobo Zhang; Yaonan Wang; Nanshan Zhong; Jianxing He
Journal:  J Thorac Dis       Date:  2020-03       Impact factor: 3.005

8.  On the accuracy of ARIMA based prediction of COVID-19 spread.

Authors:  Haneen Alabdulrazzaq; Mohammed N Alenezi; Yasmeen Rawajfih; Bareeq A Alghannam; Abeer A Al-Hassan; Fawaz S Al-Anzi
Journal:  Results Phys       Date:  2021-07-15       Impact factor: 4.476

9.  COUnty aggRegation mixup AuGmEntation (COURAGE) COVID-19 prediction.

Authors:  Siawpeng Er; Shihao Yang; Tuo Zhao
Journal:  Sci Rep       Date:  2021-07-12       Impact factor: 4.379

View more
  2 in total

1.  A Novel Approach on Deep Learning-Based Decision Support System Applying Multiple Output LSTM-Autoencoder: Focusing on Identifying Variations by PHSMs' Effect over COVID-19 Pandemic.

Authors:  Yong-Ju Jang; Min-Seung Kim; Chan-Ho Lee; Ji-Hye Choi; Jeong-Hee Lee; Sun-Hong Lee; Tae-Eung Sung
Journal:  Int J Environ Res Public Health       Date:  2022-06-01       Impact factor: 4.614

2.  In the Seeking of Association between Air Pollutant and COVID-19 Confirmed Cases Using Deep Learning.

Authors:  Yu-Tse Tsan; Endah Kristiani; Po-Yu Liu; Wei-Min Chu; Chao-Tung Yang
Journal:  Int J Environ Res Public Health       Date:  2022-05-24       Impact factor: 4.614

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.