Literature DB >> 33996401

Early prediction of coronavirus disease epidemic severity in the contiguous United States based on deep learning.

I-Hsi Kao1, Jau-Woei Perng1.   

Abstract

In November 2019, the coronavirus disease outbreak began, caused by the novel severe acute respiratory syndrome coronavirus 2. In just over two months, the unprecedented rapid spread resulted in more than 10,000 confirmed cases worldwide. This study predicted the infectious spread of coronavirus disease in the contiguous United States using a convolutional autoencoder with long short-term memory and compared its predictive performance with that of the convolutional autoencoder without long short-term memory. The epidemic data were obtained from the World Health Organization and the US Centers for Disease Control and Prevention from January 1st to April 6th, 2020. We used data from the first 366,607 confirmed cases in the United States. In this study, the data from the Centers for Disease Control and Prevention were gridded by latitude and longitude and the grids were categorized into six epidemic levels based on the number of confirmed cases. The input of the convolutional autoencoder with long short-term memory was the distribution of confirmed cases 14 days before, whereas the output was the distribution of confirmed cases 7 days after the date of testing. The mean square error in this model was 1.664, the peak signal-to-noise ratio was 55.699, and the structural similarity index was 0.99, which were better than those of the corresponding results of the convolutional autoencoder. These results showed that the convolutional autoencoder with long short-term memory effectively and reliably predicted the spread of infectious disease in the contiguous United States.
© 2021 The Authors. Published by Elsevier B.V.

Entities:  

Keywords:  Autoencoder; Convolutional neural networks; Coronavirus disease; Deep learning; Epidemic; Long short-term memory; Machine learning; Prediction; Severe acute respiratory syndrome coronavirus 2; Transmission

Year:  2021        PMID: 33996401      PMCID: PMC8105308          DOI: 10.1016/j.rinp.2021.104287

Source DB:  PubMed          Journal:  Results Phys        ISSN: 2211-3797            Impact factor:   4.476


Introduction

In November 2019, the first case of unexplained pneumonia was recorded in Wuhan, China, which soon resulted in the coronavirus disease (COVID-19) pandemic caused by the novel severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) [1]. In just over four months, its unprecedented and rapid worldwide spread affected more than 3 million people. Although the World Health Organization (WHO) declared the COVID-19 outbreak a public health emergency of international concern as early as January 2020 [2], the outbreak did not become serious in the United State until March. On March 10th, thousands of confirmed cases were recorded, and all states reported confirmed cases by March 17th. More than 10,000confirmed cases nationwide were reported by March 19th and thousands of deaths by March 26th. Subsequently, on March 27th, the numbers surpassed the public figures of China and Italy and the US became the country with the worst outbreak. COVID-19 is deadly and highly infectious. Prediction of the number and distribution of cases can allow the government to allocate medical resources in advance or to pre-deploy personnel to areas predicted to be strongly impacted by the pandemic. The rapid spread of the COVID-19 pandemic may be cause by a variety of factors, including the lack of information and transparency in the early stages of the pandemic. The timely and accurate release of epidemic data is extremely important for epidemic containment measures. In a recent study, Chinazzi et al. used a model of global metapopulation disease transmission to project the impact of travel limitations on the national and international transmission of COVID-19 [3]. The authors found that even following the travel ban in Wuhan, which was initiated on January 23rd, 2020, most Chinese cities received several infected travelers. While the quarantine of Wuhan delayed the epidemic in mainland China by three to five days, the impact at the international level was more pronounced. By mid-February, the number of cases in China decreased by nearly 80%. The authors also highlighted that early detection, hand washing, self-isolation, and family isolation may have been more effective than travel restrictions in alleviating this pandemic. Kraemer et al. used human mobility data to explain the spatial distribution of COVID-19 cases in China [4]. Following the implementation of control measures, this correlation between human mobility data and the spatial distribution of COVID-19 cases declined, with negative growth rates in most regions. Despite the demographic changes in the reported figures, they still indicate the presence of a chain of transmission outside Wuhan. The authors showed that strict control measures implemented in China greatly reduced the spread of COVID-19. In epidemiology and medical fields, mathematical modeling is a well-established method of prediction [5], [6], [7]. The Susceptible-Infected-Removed (SIR) model is one of the most widely used model [8] and has evolved to include many complex variations. Among them, are some SIR models that consider the age of the population and the prevalence of cases to improve the accuracy of the model [9]. The deformation of the SIR model has been proven by various studies to improve the effectiveness of the prediction [9], [10], [11], [12]. Some researchers have combined dynamic models with machine learning methods to achieve breakthroughs in infectious disease prediction [13], [14], [15]. Another recent study used the SIR model to predict the COVID-19 outbreak in mainland China [16]. The results of the study showed that in the initial research, because of the lack of large-scale reliable data, it was difficult to perform effective predictions using SIR models. In the absence of epidemiological data, it is difficult to objectively determine all parameters. Too many unknown parameters cause great uncertainty in model prediction. Therefore, the authors developed a simple SIR model and proved that it could effectively predict the infectivity rate of severe acute respiratory syndrome (SARS) and further used the model to predict the infectivity rate of COVID-19. However, the author also concluded that in the complex SIR model, the data provided to the model should be more accurate. Therefore, unreliable data were manually deleted to increase the reliability of the prediction model. The author reported that the more complex the model, the more accurate the prediction. However, unreasonable data may result in adverse results. By eliminating unreliable data via objective analysis, the author provided epidemic predictions under different scenarios with respect to different-level anti-epidemic measures and medical care represented by the two model parameters: infection rate and removal rate. The predictions were intended to guide decision-making in coping with ongoing SARS-CoV-2 transmission in China. However, even if the SIR model is used to predict the infectivity of COVID-19, manual intervention remains an issue. It is undesirable to manually adjust the parameters or eliminate unreliable data. In particular, when researchers do not understand local epidemic prevention policies or data substitution, it is difficult to select data accurately. Deep learning can reduce the manual manipulation of data as, in this method, the importance of each dataset is automatically managed by neural networks. Several studies have shown that deep learning can solve previously challenging issues in the fields of engineering [17], manufacturing [18], [19], [20], transportation [21], [22], [23], agriculture [24], [25], [26], and medicine [27], [28], [29], [30]. In deep learning, a variety of different architectures have been developed, such as convolutional neural networks (CNNs) [31] that are commonly used in image processing, recurrent neural networks (RNNs) [32] that are commonly used in time series, and long short-term memory (LSTM) [33] that are an improvement on RNNs. In their study, Zhang et al. collected viral RNA sequences from the Global Initiative on Sharing All Influenza Data (GISAID) and applied deep learning methods to avoid molecular dynamics simulation [34]. The authors presented a CNN structure to identify potential drugs to inhibit SARS-CoV-2 protease by performing drug screening in four chemical compound databases, showing that the deep learning model was more accurate and suitable to ensure binding tests. Zhu et al. also used deep learning to predict the host and infectivity rates of COVID-19 [35]. The authors introduced a virus host prediction method based on the fact that COVID-19 has a similar infectivity as other human coronaviruses. According to their prediction model, SARS-CoV-2 exhibited more similarity with the bat coronavirus than with coronaviruses infecting other vertebrates. Additionally, by comparing the infectivity patterns of all virus hosts among vertebrates, mink viruses showed the closest host infectivity pattern to SARS-CoV-2. Fong et al. reported the results of a study using composite Monte Carlo simulations through deep learning networks that focused on solving the problem of incomplete data available for stochastic decision making during the COVID-19 pandemic. Instead of applying the simplistic and uniform assumptions for a Monte Carlo simulation, which is a common practice, a deep learning-based composite Monte Carlo simulation was used in conjunction with fuzzy rule induction techniques [36]. Currently, several deep learning methods are being used to study COVID-19. Sajadi et al. used ERA5 to conduct climate analysis on areas severely infected with COVID-19 [37]. The results showed that, under limited latitudes, temperatures, and humidity, the distribution of major community outbreaks was consistent with the behavior of seasonal respiratory viruses. In addition, the authors proposed a simplified model that showed an increased risk of COVID-19 propagation in the area. Using weather models, it is possible to predict high-risk areas likely to be severely affected by community COVID-19 outbreaks in the next few weeks to allow public health authorities to focus on surveillance and containment in these areas. Therefore, climate is believed to affect COVID-19 transmission. The present study proposes a model based on the latitude and longitude of an area to predict the severity of the epidemic. Regardless of the traditional array processing method, the latitude and longitude data processing method can retain more data and reduce data dimensionality. Feature compression via latitude and longitude can largely preserve the climate information. After the data were pre-processed, the prediction of infectivity was conducted through a deep learning architecture proposed in this study, which was a hybrid architecture including CNN, LSTM, and an autoencoder; the convolutional autoencoder with long short-term memory is termed AL-CNN. The results showed that the AL-CNN effectively predicted the infectious disease spread of COVID-19. Previous studies have generally only predicted the total number of infected individuals; relatively few have predicted the number of infected individuals in each area. The remainder of this paper is organized as follows. In Section II, the experimental data from the WHO and US Centers for Disease Control and Prevention (CDC) are presented, along with data pre-processing. Section III presents the deep learning architectures, including AL-CNN and the convolutional autoencoder (CAE). Section IV presents the experimental results, along with a comparison of AL-CNN and CAE. Feature extraction was compared between the AL-CNN and CAE. The computational system is also discussed in this section. Section V presents the conclusions and discusses topics for future research.

Data & pre-processing

Data

The data used in this study were collected by the Johns Hopkins University Center for Systems Science and Engineering (JHU CSSE). The data were also supported by the ESRI Living Atlas Team and the Johns Hopkins University Applied Physics Lab (JHU APL) [38]. The number of confirmed cases of COVID-19, deaths, and rehabilitations were recorded. This study processed data on the number of confirmed COVID-19 case in the contiguous US. We also used another COVID-19 dataset collected by Ustayev et al. [39] based on the dataset published by the JHU CSSE and transformed it into a time series. The data were in comma-separated values (CSV) format and were updated daily. The data used in this study were from 2020.01.22 to 2020.04.06. We used data from the first 366,607 confirmed cases in the US. During this time, the number of global confirmed cases, except for China, increased from 555 to 1,345,048. By the end of the time period, the US had become one of the most severely affected countries. Additionally, the dataset included records of COVID-related deaths. The number of deaths in the US increased from 0 to 10,783 between January 22nd and April 6th, 2020. During this time, the number of deaths globally, except for China, increased from 17 to 74,565. However, we did not use data on the number of deaths in this study. Therefore, all data related to deaths were exclude. This dataset not only included the daily numbers of COVID-19 diagnoses in the US but also the location of each diagnosed person. The position accuracy in the latitude and longitude was up to the fifth decimal place. This study aimed to predict the increase in the number of COVID-19 diagnoses in an area according to its position. Recent literature indicates that COVID-19 incidence may be related to climate [37]; thus, this study considered the latitude and longitude of the affected areas to be important. Therefore, the data extracted from the database in this study included the time, location, and number of COVID-19 cases. The experimental area did not include Hawaii or Alaska.

Pre-processing

In this study, each region was categorized into six levels according to the number of COVID-19 cases. The area epidemic levels are shown in Table 1 . An epidemic level of 0 implied 0 confirmed COVID-19 cases, level 1 implied ≥ 1 and < 100 confirmed cases, level 2 implied ≥ 100 and < 1000 confirmed cases, level 3 implied ≥ 1000 and < 5000 confirmed cases, level 4 implied ≥ 5000 and < 10,000, and level 5 implied ≥ 10,000 confirmed cases.
Table 1

Area Epidemic Level.

LevelsNumber of confirmed COVID-19 cases
Level 00
Level 11N<100
Level 2100N<1000
Level 31000N<5000
Level 45000N<10000
Level 510000N

N, number of confirmed cases in the area

Area Epidemic Level. N, number of confirmed cases in the area Our model does not predict the exact number of cases in each region; rather, it predicts epidemic levels in each region. In other words, this study focused on the prediction of the area-wise severity of the epidemic. The area considered in this study was the contiguous US (Hawaii, and Alaska were excluded), which corresponded to 25° to 49° north latitude and 70° to 130° west longitude. During pre-processing, the total area was divided into 256 equal grids along the longitude, and 128 equal grids along the latitude, after which the contiguous US was divided into 32,768 global positioning system (GPS) regions (Fig. 1 ). The number of patients diagnosed with COVID-19 was added to each region and the epidemic level was graded as shown in Table 1.
Fig. 1

The contiguous United States was divided into 32,768 GPS regions.

The contiguous United States was divided into 32,768 GPS regions. After estimating the epidemic level in each GPS region, a heatmap of the epidemic can be drawn (Fig. 2 ). Fig. 2 shows the epidemic situation on April 6th, 2020, corresponding to the end of the study period. Both the input and output of the AL-CNN and CAE were heatmaps of the GPS regions. The background map was not entered into the neural network. This background image was not completely aligned with the GPS regions. Some distance errors were noted because of the curved surface of the Earth. However, the background image was only used for display purposes, and there were no errors in the actual distances in the GPS regions. Therefore, the experimental results were not affected by the non-alignment between the background image and GPS regions.
Fig. 2

Schematic diagram of epidemic level and GPS regions displayed in the contiguous United States on 2020.04.06.

Schematic diagram of epidemic level and GPS regions displayed in the contiguous United States on 2020.04.06. While visualizing the features, the value of the epidemic level was scaled from 0 to 5 to 0–255. In the training process, the min–max scaling was used to compress 0–6 levels to 0–1. The formula for the min–max scaling is as follows:where is the conversion feature and is the original feature. After the neural networks, the features are scaled back to 0–255 as an image to calculate the indicators and for visualization.

Deep learning models

We propose two architectures to predict the number of confirmed COVID-19 cases: CAE and AL-CNN. The CAE does not include the LSTM network architecture and lacks the prediction effect for the time series. However, the AL-CNN includes an LSTM to improve the prediction effect of the time series. This difference resulted in an improved prediction effect for the AL-CNN. In the following section, we introduce the network architecture and detailed design methods of the CAE and AL-CNN.

Convolutional autoencoder

Several studies have applied CAEs and demonstrated good performance in several research areas [40], [41], [42], [43]. A CAE is a neural network that uses a CNN as the core neuron and has an autoencoder (AE) architecture. This allows the CAE to possess a similar ability as CNNs in processing images. Because its architecture includes an AE, the CAE can also reconstruct images. The architecture of the CAE in this study is shown in Fig. 3 and can be roughly divided into an encoder and a decoder. The encoder included five CNN layers and five max-pooling layers, whereas the decoder included six CNN layers and five up-sampling layers. The numbers of convolutional neurons in the CNN layers were as follows: 256 in the first, second, ninth, and tenth layers; 512 in the third, fourth, seventh, and eighth layers; and 1024 in the fifth and sixth layers. The last CNN layer included seven convolutional neurons. The sizes of all convolutional neurons were .
Fig. 3

The CAE architecture in this paper.

The CAE architecture in this paper. The formula for the convolutional neurons is as follows:where is the pixel value of the convolutional neuron output, h is the maximum size of the convolutional filter, m and n are the calculated positions in the convolutional filter, is the convolutional filter of position , is the pixel value of the input, and is the bias of the convolutional filter. In the CNN layers, batch normalization was performed to avoid the vanishing gradient problem. The formula for batch normalization is as follows:where is the output value of batch normalization, is the input of the batch normalization, is the mean of the batch, h is the batch size, is the variance of the batch, and is added in the denominator for numerical stability and is an arbitrarily small constant. Following normalization, a transformation step was performed as follows:where is the weight, is the bias, and is the activation function. Except for the last layer, the activation function in all of the layers in the CAE is the rectified linear unit (ReLU), as follows: The activation function of the last layer was sigmoid, as follows: In the encoder process, after every CNN layer, a max-pooling layer was connected to reduce the image size by 50%. The formula for max-pooling is as follows:where is the pixel value of the max-pooling output and is the input of the max-pooling. In the decoder process, after every CNN layer, an up-sampling layer was connected to zoom the image to double its size. The CAE optimizer was Adam [44]. The Adam equations are as follows:where is the gradient of time , is the updated parameter, and is the stochastic objective function. Then, the biased first moment estimate and biased second raw moment estimate can be updated as follows: where and are the exponential decay rates for the moment estimates. After updating the biased first moment estimate and biased second raw moment estimate, the bias-corrected first moment estimate and bias-corrected second raw moment estimate can be computed as follows: Finally, the parameter can be updated as follows:where is the learning rate and is an extremely small value to avoid calculation error. In this study, is set to 0.001, is set to 0.9, is set to 0.999, and is set to . The loss is the mean squared error (MSE). The formula for MSE is as follows:where is the label of the image and is the predicted image. The input size of the CAE was , which represented the epidemic level 14 days prior to the test. The output size of the CAE was , which showed the predicted epidemic level 7 days following the test. The encoder reduced the image size to . There were 458,752 () neurons in the input of the encoder, whereas there were 8192 () neurons in the output. Finally, the decoder rebuilt the image with 8192 neurons into an image with 229,376 neurons. Thus, the CAE could predict the epidemic level after seven days through a large number of dimensionality reduction procedures. The results show that the CAE has a certain predictive ability, but the effect is not as good as that of the AL-CNN. The experimental results are discussed in the next section.

Convolutional autoencoder with long short-term memory

The AL-CNN is the deep-learning architecture evaluated in this study. The structure of the AL-CNN is shown in Fig. 4 . The AL-CNN and CAE differ in that the AL-CNN has the structure of an LSTM between the encoder and decoder. After the encoder, the neurons were reshaped into seven time-series steps in the AL-CNN. The last layer of convolutional neurons in the encoder process comprises 1024 neurons in the CAE. However, the neuronal output of the encoder must be a multiple of seven. Therefore, the last layer of the convolutional neurons in the encoder process was 448 in the AL-CNN.
Fig. 4

The AL-CNN structure in this paper.

The AL-CNN structure in this paper. The output of the encoder was in size, which can be reshaped to . The next step was to design the LSTM layers. LSTM performs data operations through three cores: input, output, and forget gates. The formulas for the three gates are as follows: where , , and are vectors of the input, output, and forget gates, respectively; w is the weight of the input vector in time step t; u is the weight of the previous hidden layer in time step ; x is the input vector in time step t, h is the previous hidden layer in time step ; is the bias; and is the recurrent activation function. In the AL-CNN, the recurrent activation function is a hard sigmoid function [45]. The formula for the hard-sigmoid is as follows: The output of the hard-sigmoid function was in the range of 0 to 1, which represented the degree of massage allowed to pass through the gate. The input and forget gates determined the level of data retention and the removal of data from the new memory cell. The updating process of the memory cell was calculated as follows:where is the memory state in time step t, and is the memory state in time step . The output of the current state is as follows:which is the multiplication of the memory state and output gate In the AL-CNN, there was one LSTM layer with 512 units. After processing in the LSTM layer, the data sizes were . To enable the output of the LSTM layer to be used as the input of the decoder, a reshaping layer was added after all the LSTM layers. According to the reshaping layer, the output data size was and was reshaped into a data size of . Therefore, the input size of the decoder was , whereas the output size of the decoder was , which represented the predicted epidemic level 7 days later. The AL-CNN optimizer was Adam with a learning rate of 0.001, of 0.9, and of 0.999. The loss of the AL-CNN was the MSE. Except for the last CNN layer and LSTM layers, the activation function of all layers was ReLU. The activation function of the last layer was sigmoid. The experimental results showed that the AL-CNN can predict the epidemic level of COVID-19. The results are presented in the next section.

Experimental

During the training process, early stopping was applied to avoid overfitting. When the training error was not updated for 10,000 epochs, the training process was stopped and the model weight and bias were frozen.

Experimental results

For comparison, the peak signal-to-noise ratio (PSNR) [46], MSE, and structural similarity index (SSIM) [47] were used as indicators for evaluating the rebuild heatmap. The formula for PSNR is as follows:where is the target image and is the reconstructed image. The formula for SSIM is as follows:where is the average of the signal , is the average of the signal , is the variance of the signal , is the variance of signal , and is the covariance of and . The two variables and are to stabilize the division with a weak denominator, which is shown below: where L is the dynamic range of the pixel values, and by default. These three indicators were used to calculate the degree of image restoration, in which the higher the degree of restoration, the lower the two values. The prediction results were divided into seven days to assess the prediction ability. The results obtained were the average values of each indicator for each forecast date. The average results, standard deviation errors, and confidence intervals are shown in Table 2, Table 3, Table 4 . More comparisons were performed to increase the reliability of the experiments. Therefore, the same hyperparameters were used to conduct an additional 36 experiments
Table 2

Mean Results of AL-CNN and CAE for 36 Experiments.

Prediction DateAL-CNN
CAE
Train
Test
Train
Test
MSEPSNRSSIMMSEPSNRSSIMMSEPSNRSSIMMSEPSNRSSIM
Day 11.34355.6530.9951.48256.6940.9921.04560.6250.9961.55057.2600.991
Day 21.40656.0280.9951.36257.2920.9921.12059.9950.9961.65657.1080.991
Day 31.47255.5210.9951.41057.1480.9921.11759.0060.9961.80255.4440.990
Day 41.56454.8530.9941.59256.2370.9911.22057.4360.9961.92155.3220.990
Day 51.52754.2780.9951.65955.5630.9901.25056.3740.9952.04554.2470.989
Day 61.56053.5330.9951.72054.3090.9901.24255.4030.9952.33152.9440.986
Day 71.68352.4230.9942.42352.6500.9861.58453.7680.9943.41750.8230.980
Average1.50854.6130.9951.66455.6990.9901.22557.5150.9952.10354.7350.988
Table 3

Standard Deviation Errors in the AL-CNN and CAE for 36 experiments.

Prediction DateAL-CNN
CAE
Train
Test
Train
Test
MSEPSNRSSIMMSEPSNRSSIMMSEPSNRSSIMMSEPSNRSSIM
Day 11.65911.9280.0052.87711.6120.0111.33312.8700.0042.68713.4270.012
Day 21.68312.6840.0052.64111.9630.0111.38712.7400.0042.88513.3200.012
Day 31.74512.4750.0052.34912.7030.0101.35512.0430.0042.86412.7530.011
Day 41.79012.2670.0052.42212.9050.0101.42711.5060.0042.73913.3560.012
Day 51.70911.6830.0052.42112.6510.0101.41211.1470.0042.85013.0050.011
Day 61.69510.9780.0052.27611.9260.0111.35210.6640.0042.79812.7200.014
Day 71.83610.2150.0043.15611.7730.0141.6609.8850.0043.79112.4140.019
Average1.73111.7470.0052.59212.2190.0111.41811.5510.0042.94512.9990.013
Table 4

95% Confidence Intervals of the AL-CNN and CAE for 36 Experiments.

Prediction DateAL-CNN
CAE
Train
Test
Train
Test
MSEPSNRSSIMMSEPSNRSSIMMSEPSNRSSIMMSEPSNRSSIM
Day 11.163–1.32855.177–56.4070.993–0.9961.152–1.54255.811–57.5900.987–0.9941.489–3.38959.334–60.7730.986–0.9921.245–3.18055.643–57.7840.979–0.989
Day 21.265–1.43855.128–56.4180.993–0.9961.118–1.50156.175–57.9930.988–0.9941.631–3.72859.064–60.5140.985–0.9921.351–3.46355.829–58.0380.898–0.988
Day 31.296–1.47254.779–56.0490.993–0.9961.213–1.56455.296–57.1800.987–0.9941.764–4.02258.125–59.5220.984–0.9911.550–3.81654.354–56.5150.976–0.986
Day 41.364–1.54154.323–55.5980.993–0.9961.305–1.65355.055–56.9920.987–0.9931.947–4.45656.606–57.9070.986–0.9921.662–4.10553.529–55.6560.978–0.987
Day 51.383–1.55853.674–54.8980.993–0.9961.388–1.74654.566–56.4930.986–0.9932.099–4.77055.239–56.4540.984–0.9911.732–4.33452.846–54.8860.977–0.986
Day 61.403–1.57653.078–54.2380.993–0.9961.487–1.82453.290–55.1470.985–0.9922.292–5.21454.356–55.5290.985–0.9922.120–4.90951.439–53.4490.975–0.984
Day 71.512–1.70452.041–53.1180.992–0.9952.190–2.67251.634–53.4900.981–0.9882.559–5.46053.163–54.2680.973–0.9843.270–6.02549.426–51.4020.956–0.972
Average1.341–1.51754.029–55.2470.993–0.9961.408–1.78654.547–56.4120.986–0.9921.969–4.43456.555–57.8520.983–0.9911.847–4.26253.295–55.3900.974–0.985
Mean Results of AL-CNN and CAE for 36 Experiments. Standard Deviation Errors in the AL-CNN and CAE for 36 experiments. 95% Confidence Intervals of the AL-CNN and CAE for 36 Experiments. Table 2 shows the mean results of the indicators separated by the prediction days. The average results for the seven days are also shown. When the prediction time was short, the prediction accuracies of the AL-CNN and CAE did not differ significantly. However, when predicting a longer time, the effect of the AL-CNN was better than that of the CAE. Comparisons of MSE, PSNR, or SSIM showed that the AL-CNN was better than the CAE. The training results of the CAE were better than those of the AL-CNN. However, the testing results of the CAE were worse than those of the AL-CNN. This is because the CAE causes more overfitting than the AL-CNN. Although the CAE can show lower results through training, overfitting leads to unsatisfactory test results. All three indicators displayed the same phenomenon, proving the reliability of the experiment. Table 3 lists the standard deviation errors. The standard deviation of the AL-CNN was mostly smaller than that of the CAE. The stability of the AL-CNN is better than that of the CAE. The standard deviation error was not correlated with the prediction date. Regardless of whether a closer or farther date was predicted, the value of the standard deviation error was irrelevant. Only on the seventh day did the standard deviation error of the results of the two models increase. The change in the standard deviation error of the CAE at 7 days was relatively large, whereas that of the AL-CNN was relatively small. All three indicators displayed the same phenomenon, proving the reliability of the experiment. However, in the training process, the standard deviation error of the CAE was larger than that of the AL-CNN. Thus, the CAE showed more overfitting than the AL-CNN. Table 4 presents the confidence-interval (95%) results. Evidently, the confidence interval of the AL-CNN result is smaller than that of the CAE result. Additionally, the AL-CNN was more stable than the CAE. The 95% confidence interval on the MSE gradually increased with the number of prediction days. The 95% confidence intervals for PSNR and SSIM gradually decreased with the number of prediction days. The farther the prediction date, the lower the prediction accuracy. All three indicators showed the same results, proving the reliability of this experiment. The prediction statistics clearly showed a smaller error in the model for the prediction of the epidemic level at a date closer to the current date as, for humans or mathematical models, the prediction time is closer to the prediction process. Additionally, it was more difficult to predict the parameters farther away in time. As shown in the forecast Day 1 to Day 7, the MSE of the AL-CNN increased from 1.482 to 2.423, the PSNR decreased from 56.694 to 52.65, and the SSIM decreased from 0.992 to 0.986. In addition, the MSE of the CAE increased from 1.55 to 3.417, the PSNR decreased from 57.26 to 50.823, and the SSIM decreased from 0.991 to 0.98. Both models showed difficulty in long-term predictions. Comparisons of the two models showed that the primary difference between the AL-CNN and CAE was the long-term prediction ability of the AL-CNN. In the short-term prediction of epidemic levels, the results of the AL-CNN and CAE did not differ significantly. However, when the forecast date was farther away, the prediction accuracy of the AL-CNN was significantly higher than that of the CAE. Fig. 5 shows the prediction results for the last seven days. Throughout the experiment, we used COVID-19 epidemic data for a total of 75 days. The figure shows the prediction results of the last seven days of the study period; namely, Days 69 to 75. The white color represents the epidemic level, with a brighter color indicating a higher epidemic level. Visually, the results of the AL-CNN did not differ significantly from those of the CAE. Only comparisons of the various indicators shown in Table 2, Table 3, Table 4 demonstrated the differences in the results between the models. Both are excellent prediction models, but evaluation of the indicators revealed that the AL-CNN performed better than the CAE.
Fig. 5

The last 7 days of prediction results.

The last 7 days of prediction results.

Feature observation

In this section, we discuss the feature extraction capabilities of AL-CNN and CAE. To measure the feature extraction indexes, we extracted the encoder output of the CAE and the LSTM output of the AL-CNN. The extracted features of the CAE and AL-CNN are shown in Fig. 6, Fig. 7 , respectively.
Fig. 6

The feature extraction results of CAE encoder: (a) 0° viewing angle, (b) ° viewing angle, (c) ° viewing angle, and (d) ° viewing angle.

Fig. 7

The feature extraction results of AL-CNN LSTM: (a) 0° viewing angle, (b) ° viewing angle, (c) ° viewing angle, and (d) ° viewing angle.

The feature extraction results of CAE encoder: (a) 0° viewing angle, (b) ° viewing angle, (c) ° viewing angle, and (d) ° viewing angle. The feature extraction results of AL-CNN LSTM: (a) 0° viewing angle, (b) ° viewing angle, (c) ° viewing angle, and (d) ° viewing angle. The colors in Fig. 6, Fig. 7 are shown using the hue, saturation, and value (HSV) color scale, as shown in Fig. 8 . The features of the first and last epidemic prediction days were set at the left and right ends of the color spectrum, respectively.
Fig. 8

HSV color scale used in Fig. 6, Fig. 7.

HSV color scale used in Fig. 6, Fig. 7. Regardless of the output size of the neural network, principal component analysis (PCA) [48] was used for dimension reduction for easy display. To do so, the data were initially normalized; singular value decomposition (SVD) [49] was then used to calculate the eigenvalue and eigenvector of the covariation matrix. The SVD formula is as follows:where A is the feature matrix, is the unit eigenvector of , is the unit eigenvector of , and is the square of the eigenvalue of or . The feature vectors were arranged according to their numerical values, in which feature vectors with larger feature values had more important features. We selected several feature vectors to convert and perform the following calculations:where k is the dimension reduction level. In this experiment, k was set to three. In Fig. 6, the extracted features of the CAE are randomly distributed. In the first 70 days, all epidemic features were clustered together. Even though the features are confusing, they still follow certain rules. The closer the feature points to the clustered position, the lower the epidemic level and the farther from the clustered position, the higher the epidemic level. The red, yellow, and green feature points represent features with low epidemic levels. These feature points were located at the center of the cluster. The Purple and blue represent features with high epidemic levels. These features are far from clustered feature points. As shown in Fig. 7, the extracted features of the AL-CNN showed a more regular pattern than that of the CAE. The x-axis may be a criterion for determining the epidemic level. The red, yellow, and green feature points represent features of low epidemic levels, with their position on the x-axis below 0.5. Purple and blue represent features of a high-level epidemic, with the position of their x-axis being greater than 0.5. The ability of feature extraction was also positively correlated with the prediction ability of the mode such that the stronger the feature extraction ability, the stronger the prediction ability. The addition of LSTM resulted in a significant change in the CAE. Thus, the AL-CNN showed good performance in both feature extraction and epidemic-level prediction.

Computation system

The computer hardware used in this study had an Intel I7-8700k CPU, Nvidia GeForce 1080Ti, and 32 GB of RAM. In the AL-CNN model, the training and testing processes required 53 ms and 25 ms to calculate the batch size, respectively. In the CAE model, the training and testing processes required 39 ms and 18 ms to calculate the batch size, respectively. In both processes, the AL-CNN was slightly more time-consuming than the CAE. However, as mentioned earlier, AL-CNN has improved feature extraction and prediction ability compared to CAE. This study used Python as a basic programming tool and makes extensive use of Python expansion packages, including TensorFlow, Scikit-Learn, Scikit-Image, NumPy, Pandas, and OpenCV. TensorFlow was used to design and build the neural networks. Scikit-Learn and Scikit-Image were used to organize the big data and calculate various indicators. NumPy and Pandas were used for array reading and the conversion of various data. OpenCV is a technology related to image processing. All images in this study were drawn using MATLAB R2018a because of its convenient drawing function and high image quality.

Conclusion

This study attempted to predict the epidemic levels of COVID-19 in the contiguous US by using CAE and AL-CNN models. The CAE is a deep learning neural network connected by a series of CNNs, as shown in Fig. 3. Although CAE has a good predictive ability, there remains a lack of a design framework for time-series data in neural networks. The AL-CNN is a modified version of the CAE that adds an LSTM between the encoder and decoder to increase the processing power of the model for time series data, as shown in Fig. 4. The mean square error of the AL-CNN was 1.664, the PSNR was 55.699, and the structural similarity index was 0.99, which were better than the corresponding results of the CAE, as shown in Table 2. Moreover, our study proved the predictive performance of AL-CNN. Observation of the feature extraction capabilities showed that the features extracted by AL-CNN were more regular and not cluttered as in CAE, as demonstrated in Fig. 6, Fig. 7. The model structure used in this study showed very high accuracy in the prediction of epidemic levels. Hence, this forecast can effectively assist the country in adjusting its epidemic prevention policy. This model can provide more information for observing the regional diffusion responses. Previous studies have only predicted the total number of infected individuals; relatively few have predicted the number of infected individuals in each area. The levels of the epidemic can be refined as needed. For example, when the government needs to predict whether a region will reach 1000 confirmed cases in the near future, an epidemic level should be settled with a threshold of 1000. The epidemic level can be adjusted according to the needs of the government or by the number of confirmed cases. The design of the GPS area is more complicated because it involves the design of a neural network. The present model should be practical for predicting epidemic levels in other countries or regions. Some necessary data include the confirmed case numbers for each clinic and the GPS location of each clinic. When designing a neural network, the designer must determine the appropriate size of the GPS grid. The design of the GPS grid affects the input and output of the neural network. According to the design of the GPS region, which is 32,768 in this study, the input and output size of the neural network is 14 × 128 × 256 and 7 × 128 × 256, respectively. In typical cases, the scalar of the GPS grid is expected to have a power of 2. Scalars with a power of two make the max-pooling layer easier to affect. For example, when the scalar is 16 × 16, 32 × 32, and 128 × 128, the input can be max-pooled for four, five, and seven times, respectively. By using a scalar with a power of two, the designer can change the regional scale according to their needs. In the future, we plan to organize more regional data and expand the coverage of the model. It is expected that this model can be applied not only to the contiguous US but also globally. Furthermore, additional features should be included, such as the humidity and the number of deaths. We believe that including more features can increase the prediction accuracy. Our future research will create complex deep-learning architectures that are adapted to more features.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  21 in total

Review 1.  Modeling infection transmission.

Authors:  Jim Koopman
Journal:  Annu Rev Public Health       Date:  2004       Impact factor: 21.981

2.  Strain replacement in an epidemic model with super-infection and perfect vaccination.

Authors:  Mimmo Iannelli; Maia Martcheva; Xue-Zhi Li
Journal:  Math Biosci       Date:  2005-04-26       Impact factor: 2.144

3.  Collective dynamics of 'small-world' networks.

Authors:  D J Watts; S H Strogatz
Journal:  Nature       Date:  1998-06-04       Impact factor: 49.962

4.  The pandemic of antibiotic resistance.

Authors:  R M Anderson
Journal:  Nat Med       Date:  1999-02       Impact factor: 53.440

5.  Long short-term memory.

Authors:  S Hochreiter; J Schmidhuber
Journal:  Neural Comput       Date:  1997-11-15       Impact factor: 2.026

Review 6.  Mathematical and computational challenges in population biology and ecosystems science.

Authors:  S A Levin; B Grenfell; A Hastings; A S Perelson
Journal:  Science       Date:  1997-01-17       Impact factor: 47.728

7.  Mapping the transmission risk of Zika virus using machine learning models.

Authors:  Dong Jiang; Mengmeng Hao; Fangyu Ding; Jingying Fu; Meng Li
Journal:  Acta Trop       Date:  2018-06-19       Impact factor: 3.112

8.  Modeling Task fMRI Data Via Deep Convolutional Autoencoder.

Authors:  Heng Huang; Xintao Hu; Yu Zhao; Milad Makkie; Qinglin Dong; Shijie Zhao; Lei Guo; Tianming Liu
Journal:  IEEE Trans Med Imaging       Date:  2017-06-15       Impact factor: 10.048

9.  Early Prediction of the 2019 Novel Coronavirus Outbreak in the Mainland China Based on Simple Mathematical Model.

Authors:  Linhao Zhong; Lin Mu; Jing Li; Jiaying Wang; Zhe Yin; Darong Liu
Journal:  IEEE Access       Date:  2020-03-09       Impact factor: 3.367

10.  Deep Learning Based Drug Screening for Novel Coronavirus 2019-nCov.

Authors:  Haiping Zhang; Konda Mani Saravanan; Yang Yang; Md Tofazzal Hossain; Junxin Li; Xiaohu Ren; Yi Pan; Yanjie Wei
Journal:  Interdiscip Sci       Date:  2020-06-01       Impact factor: 2.233

View more
  1 in total

1.  Comparison of Eye and Face Features on Drowsiness Analysis.

Authors:  I-Hsi Kao; Ching-Yao Chan
Journal:  Sensors (Basel)       Date:  2022-08-30       Impact factor: 3.847

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.