Literature DB >> 19683069

Comparing early outbreak detection algorithms based on their optimized parameter values.

Xiaoli Wang1, Daniel Zeng, Holly Seale, Su Li, He Cheng, Rongsheng Luan, Xiong He, Xinghuo Pang, Xiangfeng Dou, Quanyi Wang.   

Abstract

BACKGROUND: Many researchers have evaluated the performance of outbreak detection algorithms with recommended parameter values. However, the influence of parameter values on algorithm performance is often ignored.
METHODS: Based on reported case counts of bacillary dysentery from 2005 to 2007 in Beijing, semi-synthetic datasets containing outbreak signals were simulated to evaluate the performance of five outbreak detection algorithms. Parameters' values were optimized prior to the evaluation.
RESULTS: Differences in performances were observed as parameter values changed. Of the five algorithms, space-time permutation scan statistics had a specificity of 99.9% and a detection time of less than half a day. The exponential weighted moving average exhibited the shortest detection time of 0.1 day, while the modified C1, C2 and C3 exhibited a detection time of close to one day.
CONCLUSION: The performance of these algorithms has a correlation to their parameter values, which may affect the performance evaluation.

Entities:  

Mesh:

Year:  2009        PMID: 19683069      PMCID: PMC7185865          DOI: 10.1016/j.jbi.2009.08.003

Source DB:  PubMed          Journal:  J Biomed Inform        ISSN: 1532-0464            Impact factor:   6.317


Introduction

Following the outbreak of severe acute respiratory syndrome [1] in 2003, there has been a growing recognition of the necessity and urgency of early outbreak detection of infectious diseases. In January 2004, the National Disease Surveillance, Reporting and Management System were launched in China. The system which covers 37 infectious diseases has the potential to provide timely analysis and early detection of outbreaks. However, as the passive surveillance system relies on accumulated case and laboratory reports, which are often delayed and sometimes incomplete, the opportunity to contain the spread of the disease is often missed. As increasing numbers of early outbreak detection algorithms are now being used in public health surveillance [2], [3], [4], [5], [6], [7], [8], [9], there is a need to evaluate their performance. Due to a lack of complete and real data pertaining from historical outbreaks, the performance of these systems have been previously difficult to evaluate [10]. Adding to these difficulties is the fact that the information obtained from historical outbreaks may be heterogeneous, due to changes in the outbreak surveillance criteria’s over time. In order to compensate for missing or heterogeneous information, semi-synthetic datasets can be created which contain the outbreak signals, using a software tool. By using this tool, the parameters of the outbreak including the desired duration, temporal pattern and the magnitude (based on a predefined criteria), can be specially set. This approach has been documented in a number of previous studies, which have compared the performance of early outbreak detection algorithms using simulated outbreaks [11], [12], [13], [14], [15], [16], [17], [18]. The simulation enables the performance assessment and provides much-need comparative findings about outbreak detection algorithms. However, there are still limited studies examining how the performance varies with the values of these algorithm parameters. Our study aimed to observe the relationship between the algorithms’ performance and their parameters values. The outcomes of this study may help improve the accuracy and objectivity of the evaluation of these algorithms and provide guidelines for future research and implementation.

Methods

Baseline data

Bacillary dysentery is one of the key epidemic potential diseases in Beijing. It commonly occurs in summer and in regions with high population densities. With economic development and improvements in sanitary conditions in China, the incidence of bacillary dysentery has decreased substantially from 1990 to 2003 [19]. Between 2004 and 2007, data from the National Disease Surveillance Reporting and Management System showed that the average incidence rate of bacillary dysentery was 235.9 cases per 100,000 in Beijing. Whilst there has been a substantial decline in the disease burden, bacillary dysentery continues to be a major public health problem in Beijing. The observed daily case counts of bacillary dysentery from 2005 to 2007 in Beijing were extracted from the National Disease Surveillance Reporting and Management System [20] for this study. The onset date of illness and area code at the sub-district level was extracted for each reported case. This data was used as the baseline for the outbreak simulation. Data from 2005 to 2006 were used to adjust and optimize the parameter values of the algorithms, while data from 2007 were used to evaluate the algorithms.

Data simulation

The outbreak criteria was defined on the basis of the bacillary dysentery reporting criteria specified in the National Protocol for Information Reporting and Management of Public Health Emergencies (Trial) [21]. This protocol was issued by the Health Emergency Office of the Ministry of Health (MOH) at the end of 2005. In the protocol, a bacillary dysentery outbreak was defined as the occurrence of 10 or more bacillary dysentery cases in the same school, natural village or community within 1–3 days. Based on this definition, there was only one actual outbreak in the summer of 2007. During this outbreak, 10 children from a middle school were clinically diagnosed as having bacillary dysentery and four were culture positive for Shigella sonnei. The first case became ill on the evening of the 21st of July and was taken to hospital the next day. Two cases were reported on the 22nd of July, a further four on the 23rd of July, and two on the 24th July. As there were insufficient documents collected during the outbreak, a simulated outbreak signal had to be produced. Before the simulation, the actual outbreak was excluded by replacing actual data with a 7-day moving average for fear of contamination. Our simulation approach used semi-synthetic data, that was, authentic baseline data injected with artificial signals [9]. The AEGIS-Cluster Creation Tool (AEGIS-CCT) was used to generate outbreak signals [22]. First, the duration was fixed at three days and the outbreak magnitude varied from 10 to 20. The outbreak magnitude was fixed at 10 cases and the duration was varied from one to three days. The temporal progression of these outbreaks included a random, a linear, and an exponential growth spread (12 signals for each temporal progression pattern). A total of 36 different outbreak signals were finally simulated. Considering the spatial distribution and seasonal variability of bacillary dysentery, we randomly selected 30 (10 for each pattern) from a possible 100 sub-districts (townships), where the incidence was higher than the average incidence in Beijing, and then randomly selected one day as the starting date of an outbreak from the high incidence seasons. The remaining six outbreak signals were randomly added to the low incidence seasons and areas. Simulations injected into the baseline data from the selected sub-districts (2005–2006) were used to observe the relationship between the algorithm performance and the parameter value. This data allowed us to select the optimal combination of parameter values. Simulations added to the baseline data from 2007 were used to evaluate the algorithm. In order to reduce sampling errors, means were calculated by repeating the sampling 50 times.

Evaluation indices

Evaluation indices included sensitivity, specificity and time to detection [14]. An outbreak was considered to be detected when a signal was triggered: (1) within the same period as the start and end date of the particular simulated outbreak; and (2) within the same sub-district as what the simulation was geographically located in. In our study, sensitivity was defined as the number of outbreaks in which ⩾1 day was flagged, divided by the number of simulated outbreaks. Specificity was defined as the number of days that were not flagged divided by the number of non-outbreak days. Time to detection was defined as the interval between the beginning of the simulated outbreak and the first day flagged by the algorithm, divided by the number of simulated outbreaks. Time to detection was zero, if the algorithm flagged a simulated outbreak on the first day. Time to detection was three, if the algorithm did not produce a flag on any of the days during the period of the simulated outbreak. Time to detection is an integrated index that reflects both timeliness and sensitivity of an algorithm.

Evaluation criteria

We intended to find a simple and practical criterion to evaluate the performance of these algorithms. Generally, the parameter values with the shortest time to detection were considered as preferable. The disparity in specificity between the parameter values was also taken into consideration. Priority was given to the value with the higher specificity, if the time to detection was either equal to or had a difference of less than half a day and the difference between the specificities was >5.0%.

Outbreak detection algorithms

We compared the performance of five outbreak detection algorithms, the exponential weighted moving average (EWMA), C1-MILD (C1), C2-MEDIUM (C2), C3-ULTRA (C3) and the space–time permutation scan statistic model. We calculated the EWMA, using a 28-day baseline based on day t  − 30 through till day t  − 3 within each sub-region [15]. If the observed values were x i  ∼  N (μ,  σ 2), the weighted daily counts of each sub-district were calculated as: In the algorithm, λ (0 <  λ  < 1) was the weighting factor, and k was the control limit coefficient [15], [23]. They are the adjustable parameters. Based on the range in values of k found in previous literature [23], k was set as 0 <  k  ⩽ 3. The adjustment interval for λ and k was set as 0.1 and 0.5, respectively. The moving standard deviation (S) was used as the estimate of σ; and the moving average (MA) was used as the estimate of μ. The cumulative sum (CUSUM) algorithm keeps track of the accumulative deviation between the observed and expected values. For CUSUM, the accumulated deviation S was defined as: S_{0} = 0·kσ is the allowed shift from the mean to the detected. S is the current CUSUM calculation, and S −1 is the previous CUSUM calculation. We found that there was an aberration when the mean μ 0 shifted to μ 0  +  kσ· h was the decision value. In EARS, k was set as 1 and when S  >  h  =  2, an alarm would be trigged [24]. When the denominator σ equals to zero, 0.2 was taken to replace zero in EARS. However, as both sides of the equation can be multiplied (S  = max (0, S −1  + ((X  − (μ 0  +  kσ))/σ )) > 2) by σ, the decision value was changed to hσ (referred to as H). BioSense originally implemented the C1, C2 and C3 methods but has since modified the C2 method (referred to as W2). In our study, we did not use the threshold; k or decision values set in EARS, rather we adjusted these values to achieve a preferable efficiency for aberration detection. Additionally, we did not use 0.2 when σ was 0, rather the actual value. Based on the previous literatures [13], [25], [26], we determined the value range of H and k, as 3σ  ⩽  H  ⩽  5σ and 0 <  k  ⩽ 1.5, respectively. The adjustment interval for k and H was set as 0.1 and 0.5σ, respectively. We modified the three original CUSUM referred to as C1, C2 and C3 to C1′, C2′ and C3′ in the reporting of the results in this study. The equation is written as C3′ was the sum of C, C and C derived from C2′. MA 1 was the moving sample average and S 1 was the moving standard deviation of the case count reported from baseline. MA 2 and S 2 were the moving sample average and moving standard deviation of the case count reported during baseline period, with a 2-day lag. The moving standard deviation (S) was used as the estimate of σ; and the moving average (MA) was used as the estimate of μ. The length of the baseline comparison period for all three methods was 7-days in order to account for the day of the week effect [13], [14]. The space–time permutation scan statistic model utilizes thousands to millions of overlapping cylinders to define the scanning window, each of which is a possible candidate for an outbreak. The circular base represents the geographical area of the potential outbreak from zero to some designated maximum value. The height of the cylinder represents the time period of a potential cluster. The probability function for any given window is proportional to [27], [28]:where C was the observed number of cases in subzone z and during day d. C was the total number of observed cases during the whole study phase T for the whole study region. C was the observed case count scanned in cylinder A. The generalized likelihood ratio (GLR) was calculated as a measure of the evidence that cylinder A contains an outbreak. Among the many cylinders evaluated, the one with the maximum GLR constitutes the space–time cluster of cases that is least likely to be a chance occurrence and, hence, is the primary candidate for a true outbreak. The size and location of the scanning window is under dynamic change [28]. The maximum temporal cluster size was determined by considering the incubation period of the disease studied. For bacillary dysentery, the average incubation period was 1–3 days. Therefore, the maximum temporal cluster size in this study was set as (1d, 3d, 5d and 7d). The maximum spatial cluster size can be determined in virtue of the geographical area or the proportion of the whole population. Since data on the proportion of the population in each sub-district were unavailable, the maximum spatial cluster size in this study was set as (2, 5, 8 and 10 km), referring to the geographical area of each sub-district. The performance was analyzed using P values of 0.05.

Data analysis

Analyses were undertaken using EXCEL, SPSS software (version 13.0 for Windows; SPSS Inc., Chicago, IL), AEGIS-CCT (available from http://sourceforge.net/projects/chipcluster/), JAVA programming (available from http://java.com/zh_CN/) and SaTScan (available from www.satscan.org). SPSS was used for data processing, descriptive statistics and the chi-square test. The Bonferroni correction was applied for multiple comparisons to control the family wise error rate. The significance level α for an individual test was calculated by dividing the family wise error rate (0.05) by the number of tests [29]. EWMA and the cumulative sum were coded by JAVA programming to find out whether the incidence level was abnormal. SaTScan was used to analyze the clustering of cases in different sub-districts in Beijing based on space–time permutation scan statistics and whether the incidence level was abnormal.

Results

Adjustment and optimization of parameter

The correlation coefficients between the three evaluation indices (sensitivity, specificity and time to detection) and parameter values were calculated. Table 1 showed the correlation coefficients with Pearson’s r and P values. All algorithms showed strong relation between the evaluation indices and the parameter’ values, except space–time permutation scan statistic. Great majority of the correlation was statistically significant, with P values less than 0.05(two-tailed). However, for space–time permutation scan statistic, specificity showed no relation to the spatial cluster size. Only when the maximum temporal cluster size was set as 3d, both sensitivity and time to detection exhibited a significant correlation with the spatial cluster size (P  < 0.05).
Table 1

Correlation between the three evaluation indices (sensitivity, specificity and time to detection) and parameter values for five algorithms with P values#.

AlgorithmsSensitivity
Specificity
Time to detection
Pearson’s rP valuePearson’s rP valuePearson’s rP value
λ value for EWMA (k = 1.0)0.5260.118−0.8970.000
λ value for EWMA(k = 2.0)0.8370.0030.6840.029−0.9900.000
λ value for EWMA (k = 3.0)0.9210.0000.7800.008−0.9900.000
k value for C1′ (H = 3σ)−0.9180.0000.9690.0000.9870.000
k value for C1′ (H = 4σ)−0.9730.0000.9370.0000.9920.000
k value for C1′ (H = 5σ)−0.9080.0000.9040.0010.9950.000
k value for C2′ (H = 3σ)−0.7590.0000.9820.0000.9730.000
k value for C2′ (H = 4σ)−0.9070.0000.9610.0000.9840.000
k value for C2′ (H = 5σ)−0.9380.0000.9430.0010.9590.000
k value for C3′ (H = 3σ)−0.6940.0040.9980.0000.9560.000
k value for C3′ (H = 4σ)−0.8180.0000.9940.0000.9620.000
k value for C3′ (H = 5σ)−0.8660.0000.9950.0000.9600.000
Spatial cluster size for space–time permutation scan statistic (3d)−0.9580.0420.9620.038
Spatial cluster size for space–time permutation scan statistic (7d)−0.8350.1650.7560.244
Spatial cluster size for space–time permutation scan statistic (10d)−0.7460.2540.5380.462

Pearson’s r was calculated directly for those variables showing linear relation. While for those variables showing relation of log linear, Pearson’s r was calculated after applying logarithmic transformation.

P value was less than 0.05 (2-tailed).

Correlation between the three evaluation indices (sensitivity, specificity and time to detection) and parameter values for five algorithms with P values#. Pearson’s r was calculated directly for those variables showing linear relation. While for those variables showing relation of log linear, Pearson’s r was calculated after applying logarithmic transformation. P value was less than 0.05 (2-tailed). Fig. 1, Fig. 2, Fig. 3, Fig. 4 describe the average sensitivity, specificity and time to detection of the five algorithms. The top plot of Fig. 1 shows the sensitivity versus λ values for the three control limit coefficients (k). In all of the combinations of λ and k values, the sensitivities were greater than 90%. As λ increased from 0 to 0.9, the sensitivity also increased. The middle plot of Fig. 1 shows the specificity of the three k values. Specificity of the three k values had a similar change trend by λ value, increasing until λ  = 0.3, and then declining gradually. The bottom plot of Fig. 1 shows the effect of λ values on detection timeliness of EWMA. Time to detection declined gradually with the increasing λ values. Among these combinations of different λ and k values, λ  = 0.9, k  = 1.0 showed the shortest detection time, with a specificity of 89.5%. There were only two combination of λ and k values that had a detection time longer than half a day (λ  = 0.1, k  = 2.0 and λ  = 0.1, k  = 3.0). Out of the remaining combinations, there were 11 which had specificity greater than 89.5%. Within these 11 combinations, λ  = 0.7, k  = 3.0 showed the greatest specificity (97.2%). According to the evaluation criteria, we concluded that λ  = 0.7, k  = 3.0 was the optimal parameter for EWMA.
Fig. 1

Sensitivity, specificity and time to detection for EWMA with a combination of λ and k (λ = 0.1, 0.2, ..., 1.0; k = 1.0, 2.0 and 3.0), shown from top to bottom.

Fig. 2

Sensitivity, specificity and time to detection for C1′ with a combination of k and H (k = 0.1, 0.2, ..., 1.5; H = 3σ, 4σ and 5σ), shown from top to bottom.

Fig. 3

Sensitivity, specificity and time to detection for C2′ with a combination of k and H (k = 0.1, 0.2, ..., 1.5; H = 3σ, 4σ and 5σ), shown from top to bottom.

Fig. 4

Sensitivity, specificity and time to detection for C3′ with a combination of k and H (k = 0.1, 0.2, ..., 1.5; H = 3σ, 4σ and 5σ), shown from top to bottom.

Sensitivity, specificity and time to detection for EWMA with a combination of λ and k (λ = 0.1, 0.2, ..., 1.0; k = 1.0, 2.0 and 3.0), shown from top to bottom. Fig. 2 shows the influence of different H and k values on sensitivity, time to detection and specificity. The sensitivity was shown to decrease as k increased. As the sensitivity decreased, time to detection increased. Among the combinations of H and k values, (H  = 3σ, k  = 0.1) had the shortest time to detection of 0.3 day (specificity: 55%). There were 14 combinations with a detection time of half a day longer than (H  = 3σ, k  = 0.1). All of these 14 combinations had specificities greater than 55%, with the highest one being 95.6%, when H  = 5σ, k  = 0.4. According to the evaluation criteria, (H  = 5σ, k  = 0.4) was found to be the optimal combination for C1′. Sensitivity, specificity and time to detection for C1′ with a combination of k and H (k = 0.1, 0.2, ..., 1.5; H = 3σ, 4σ and 5σ), shown from top to bottom. The relationship between performance and the combination of H and k values for C2′ is shown in Fig. 3 . We found that sensitivity declined as k increased from 0.1 to 1.5. In comparison, specificity and time to detection increased as sensitivity declined. The combination of (H  = 3σ, k  = 0.1) showed the shortest detection time (0.2d), with a specificity of 46.2%. Similarly, 14 combinations had a detection time which was half a day longer than (H  = 3σ, k  = 0.1). The specificities for all of these 14 combinations was greater than 46.2%, with the highest one recorded at 88.6%, when H  = 4σ, k  = 0.5. Accordingly, (H  = 4σ, k  = 0.5) was thought the optimal combination for C2′. Sensitivity, specificity and time to detection for C2′ with a combination of k and H (k = 0.1, 0.2, ..., 1.5; H = 3σ, 4σ and 5σ), shown from top to bottom. Fig. 4 shows the influence of sensitivity, time to detection and specificity of H and k values for C3′. The specificity and time to detection had an overall growth of λ value. Sensitivity declined gradually as λ increased. Among the combinations of H and k values, (H  = 3σ, k  = 0.1) had the shortest time to detection (0.1d), with a specificity of 18.3%. Likewise, there were 14 combinations with a detection time half a day longer than (H  = 3σ, k  = 0.1). 13 out of these 14 combinations had specificities greater than 18.3%, the highest one being 73.9%, when H  = 3σ and k  = 0.8. Consequently, (H  = 3σ, k  = 0.8) was thought as the optimal combination for C3′. Sensitivity, specificity and time to detection for C3′ with a combination of k and H (k = 0.1, 0.2, ..., 1.5; H = 3σ, 4σ and 5σ), shown from top to bottom. We found that the space–time permutation scan statistics exhibited no real difference in the specificity when the parameter combinations were changed (Table 2 ). When the maximum temporal cluster size was set as 3d and the maximum spatial cluster size of 2 km, the detection time was found to be the shortest. This combination also resulted in the highest specificity and sensitivity. Thus the optimal parameter was taken as 3d (maximum temporal cluster size) and 2 km (maximum spatial cluster size).
Table 2

Average sensitivity, specificity and time to detection for space–time permutation scan statistics with various combinations of spatial and temporal cluster size.

Spatial cluster size (km)Temporal cluster size = 3d
Temporal cluster size = 7d
Temporal cluster size = 10d
Sensitivity (%)Specificity (%)Time to detection (d)Sensitivity (%)Specificity (%)Time to detection (d)Sensitivity (%)Specificity (%)Time to detection (d)
296.999.90.290.699.90.496.999.90.3
584.499.90.581.399.90.681.399.90.6
884.499.90.584.499.90.587.599.90.4
1081.399.90.681.399.90.684.499.90.5
Average sensitivity, specificity and time to detection for space–time permutation scan statistics with various combinations of spatial and temporal cluster size.

Evaluation of the algorithms

Five commonly used algorithms were evaluated by comparing the performance with their optimized parameters values. The performance of these algorithms is shown in Table 3 with P values. According to Bonferroni’s procedure, the significance level α for an individual test was calculated by dividing the family wise error rate (0.05) by four. This was found to be 0.0125. Of the algorithms evaluated, space–time permutation scan statistics had a higher average specificity than any other algorithms (P  < 0.001), followed by EWMA (95.2%), while C3′ showed the lowest specificity (73.7%). EWMA had the shortest time to detection (0.1d), while C1′ showed the longest time to detection of one day. Space–time permutation scan statistics had a relatively longer time to detection compared to EWMA (0.2d), but this difference was not statistically significant (P  = 0.081 > 0.0125). According to the evaluation criteria and statistical test, we could conclude that space–time permutation scan statistics was the optimal algorithm, followed by EWMA. Space–time permutation scan statistics had a specificity of 99.9%, which meant that only one false alarm occurred per 1000 days, whereas EWMA was evaluated to trigger one false alarm for every 21 days.
Table 3

Specificity and time to detection for five algorithms based on the optimize combinations of parameter value.

AlgorithmsSpecificity (%)
Time to detection (d)
Meanχ2 Test lower CIMeanχ2 Test lower CI
EWMA (0.7, 3.0)95.2<0.0010.1>0.0125
C1′ (0.4, 5σ)95.7<0.0011.0<0.001
C2′ (0.5, 4σ)89.2<0.0010.8<0.001
C3′ (0.8, 3σ)73.7<0.0010.7<0.001
Space–time permutation scan statistic (3d, 2 km) 99.90.2

Space–time permutation scan served as control group. According to Bonferroni’s procedure, the family wise error rate was 0.05, divided by the number of test. The significance level α for an individual test was 0.05/4, being 0.0125.

Specificity and time to detection for five algorithms based on the optimize combinations of parameter value. Space–time permutation scan served as control group. According to Bonferroni’s procedure, the family wise error rate was 0.05, divided by the number of test. The significance level α for an individual test was 0.05/4, being 0.0125.

Discussion

The burden of bacillary dysentery has long been thought to be great in many developing countries [30]. Detecting outbreaks in their early stages may prevent secondary infections, and subsequently an epidemic from occurring. The benefits of this extend not only to the individual, but also to the community in terms of morbidity prevented and costs saved. From the case study in 2007, the outbreak was detected when the accumulated number of cases reached the threshold (10 cases in 3 days within the same geographic area). The problem with this method of detection is that the optimal opportunity to curb an outbreak is often missed. In the event of a pandemic influenza or another emerging inflection, missing this opportunity may have national or global implications. We observed that the effects of the same algorithm varied significantly with different parameter values. For example, the time to detection and specificity were 73.9% and 0.6d for C3′ (H  = 3σ, k  = 0.8) versus 61.8% and 0.6d for C2′ (H  = 3σ, k  = 0.2). If the performance of C3′ and C2′ were compared with these values, C3′ (H  = 3σ, k  = 0.8) seemed to be better than C2′ (H  = 3σ, k  = 0.2) according to the evaluation criteria, which might lead to the conclusion that C3′ was more effective than C2′. In fact, C2′ (H  = 4σ and k  = 0.4) had a detection time of 0.6d and a specificity of 85.8%, 11.9% higher than 73.9% (C3′, H  = 3σ, k  = 0.8). In this case, C2′ (H  = 4σ and k  = 0.4) were better than C3′ (H  = 3σ and k  = 0.8). The difference in performance of the two algorithms is largely caused by the difference between parameters’ values. Therefore, parameter values should be optimized prior to the performance evaluation of algorithms. A wide range of outbreak detection algorithms are available including: temporal, spatial and spatial–temporal [31]. In this study, we used both the temporal and spatial information of the reported cases. The temporal information refers to the onset date of the illness, and spatial information refers to the sub-district where the case currently resides at. CUSUM and EWMA are commonly used to analyze the temporal data, as they can be adjusted to identify a meaningful change from the expected range of data values. We calculated the daily case counts reported for each sub-district, and then judged whether the change from the expected value was significant within each sub-district. So in our study, CUSUM and EWMA can also give us both the temporal and spatial information of the signal. Our study focused on the correlation between algorithm parameter values and their performance. By calculating the correlation coefficient and comparing the performance of different algorithms with various values, we observed a strong correlation between them. The differences in the parameter values may have resulted from a difference in the performances among these algorithms. Consequently, we recommend that before evaluating the effectiveness of an outbreak detection algorithm, parameter values should be optimized to remove the noise which has resulted from the potential influence of parameter value for a given disease. In our study we found that space–time permutation scan statistics and the EWMA outperformed other algorithms both in terms of timeliness and accuracy for detecting bacillary dysentery outbreaks. EWMA applies weighting factors which decrease exponentially. The choice of weighting factor λ is the key for successful outbreak detection. With proper λ value, EWMA control procedure can be adjusted to be sensitive to a small or gradual drift in the process. We feel that adjusting λ value should be an imperative step before applying EWMA into practice. Space–time permutation scan statistics consider both the temporal and spatial factors. The scanning window is under dynamic change to avoid selection bias. However, space–time scan statistics do not consider population movements. In addition, space–time scan statistics can only identify clusters in simple regular shapes. If the cluster does not conform to a regular shape, the algorithm may have a poor performance. Therefore, when space–time permutation scan statistics are used to detect the outbreaks, it is imperative to understand the cluster shape. Only in the right shape, can space–time permutation scan statistics demonstrate a high detection efficacy. Aside from these limitations, the use of space–time permutation scan statistics allowed the early outbreak detection for bacillary dysentery. Previously, Hutwagner et al. [14] compared the time to detection with simulation based on influenza like illness and pneumonia data. In her study, C1, C2 and C3 were found to have an increasing time to detection. In comparison, we found a decline in the detection time for our modified C1, C2 and C3. These differences in the time to detection calculations may explain the differences between the two studies. In our study, when the algorithms failed to detect the simulated outbreak, time to detection was set as the largest value (3 days). As we know, C1, C2 and C3 have increasing sensitivities. Obviously, as the sensitivity increased from C1 to C3, the number of missed outbreaks decreased and consequently the time to detection declined accordingly. An integrated time to detection might be recommended, in order to address this limitation [14]. Theoretically, the optimal parameter value can maximize the algorithm’s ability to detect aberration in disease incidence and minimize the probability of producing a false alarm. The balance between the accuracy and timeliness is still a matter of debate. In our study, we set simple and practical evaluation criteria’s. Considering the time to detection integrating effect of sensitivity, we simplified the three evaluation indices to two, time to detection and specificity. The former reflected both the timeliness and sensitivity, and the latter reflected the accuracy of outbreak detection. We made timeliness the priority over accuracy due to bacillary dysentery’s short incubation period and the fact that it can be both food-borne and water-borne. When deciding which index should be given the priority, practitioners should take the length of incubation, the mode of transmission and the current situation (climatic, social, demographic, economic factors, etc.) into consideration. The variation in patterns of the evaluation indices with the change of parameter values observed in our study was found to be consistent with previous related studies [9], [12], [14], [15], [32], [33]. For example, Hutwagner et al. [14] observed that C1, C2 and C3 had increasing sensitivity, but a decreasing specificity as the sensitivity increased. In our study, we also observed this change in sensitivity and specificity in our modified C1′, C2′ and C3′. In our study we observed a growth in sensitivity and specificity as weighing values increased from 0 to 0.3. It seemed that the range of weighting values from 0.4 to 0.9 enabled a better performance. This recommendation was also made by Jackson et al. [15], who suggested weighing values of 0.4 and 0.9 for EWMA. There are several factors which may limit the generalization of our findings. To apply these five algorithms, information on the specific setting (workplaces, schools etc.) is often required. This information is usually not available in the current National Disease Surveillance Reporting and Management System in China. Consequently, the sensitivity of the five algorithms may be less when a bacillary dysentery outbreak occurs in a school, as the cases may be scattered in different sub-districts. It is therefore important to collect extra information on workplaces, schools and other units. Due to a lack of actual outbreaks, we injected simulated outbreaks into the baseline so we could undertake a performance assessment on these outbreak detection algorithms. We changed the size, magnitude, temporal progressive pattern, season and spatial distribution of bacillary dysentery, in order to have a variety of outbreak conditions to test. As these are approximations, it is difficult to evaluate how close our simulations came to the actual outbreak. Consequently, further research is needed in predicting the actual performance of these algorithms.
  12 in total

1.  Adjusting outbreak detection algorithms for surveillance during epidemic and non-epidemic periods.

Authors:  Zhongjie Li; Shengjie Lai; David L Buckeridge; Honglong Zhang; Yajia Lan; Weizhong Yang
Journal:  J Am Med Inform Assoc       Date:  2011-08-11       Impact factor: 4.497

2.  Epidemic features affecting the performance of outbreak detection algorithms.

Authors:  Jie Kuang; Wei Zhong Yang; Ding Lun Zhou; Zhong Jie Li; Ya Jia Lan
Journal:  BMC Public Health       Date:  2012-06-08       Impact factor: 3.295

3.  Outbreak detection algorithms for seasonal disease data: a case study using Ross River virus disease.

Authors:  Anita M Pelecanos; Peter A Ryan; Michelle L Gatton
Journal:  BMC Med Inform Decis Mak       Date:  2010-11-24       Impact factor: 2.796

4.  Hand, foot and mouth disease in China: evaluating an automated system for the detection of outbreaks.

Authors:  Zhongjie Li; Shengjie Lai; Honglong Zhang; Liping Wang; Dinglun Zhou; Jizeng Liu; Yajia Lan; Jiaqi Ma; Hongjie Yu; David L Buckeridge; Chakrarat Pittayawonganan; Archie C A Clements; Wenbiao Hu; Weizhong Yang
Journal:  Bull World Health Organ       Date:  2014-06-23       Impact factor: 9.408

5.  Evaluation and comparison of statistical methods for early temporal detection of outbreaks: A simulation-based study.

Authors:  Gabriel Bédubourg; Yann Le Strat
Journal:  PLoS One       Date:  2017-07-17       Impact factor: 3.240

6.  How to select a proper early warning threshold to detect infectious disease outbreaks based on the China infectious disease automated alert and response system (CIDARS).

Authors:  Ruiping Wang; Yonggen Jiang; Engelgau Michael; Genming Zhao
Journal:  BMC Public Health       Date:  2017-06-12       Impact factor: 3.295

7.  Implementation and Initial Analysis of a Laboratory-Based Weekly Biosurveillance System, Provence-Alpes-Côte d'Azur, France.

Authors:  Michael Huart; Gabriel Bedubourg; Cédric Abat; Philippe Colson; Jean Marc Rolain; Hervé Chaudet; Pierre Edouard Fournier; Didier Raoult; Xavier Deparis
Journal:  Emerg Infect Dis       Date:  2017-04       Impact factor: 6.883

8.  Improving the performance of outbreak detection algorithms by classifying the levels of disease incidence.

Authors:  Honglong Zhang; Shengjie Lai; Liping Wang; Dan Zhao; Dinglun Zhou; Yajia Lan; David L Buckeridge; Zhongjie Li; Weizhong Yang
Journal:  PLoS One       Date:  2013-08-19       Impact factor: 3.240

9.  Development of ClickClinica: a novel smartphone application to generate real-time global disease surveillance and clinical practice data.

Authors:  Benedict Daniel Michael; David Geleta
Journal:  BMC Med Inform Decis Mak       Date:  2013-07-02       Impact factor: 2.796

10.  Evaluation of the performance of a dengue outbreak detection tool for China.

Authors:  Honglong Zhang; Zhongjie Li; Shengjie Lai; Archie C A Clements; Liping Wang; Wenwu Yin; Hang Zhou; Hongjie Yu; Wenbiao Hu; Weizhong Yang
Journal:  PLoS One       Date:  2014-08-29       Impact factor: 3.240

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.