Literature DB >> 35239641

Estimates of the basic reproduction number for rubella using seroprevalence data and indicator-based approaches.

Timos Papadopoulos1,2, Emilia Vynnycky1,3,4.   

Abstract

The basic reproduction number (R0) of an infection determines the impact of its control. For many endemic infections, R0 is often estimated from appropriate country-specific seroprevalence data. Studies sometimes pool estimates from the same region for settings lacking seroprevalence data, but the reliability of this approach is unclear. Plausibly, indicator-based approaches could predict R0 for such settings. We calculated R0 for rubella for 98 settings and correlated its value against 66 demographic, economic, education, housing and health-related indicators. We also trained a random forest regression algorithm using these indicators as the input and R0 as the output. We used the mean-square error to compare the performances of the random forest, simple linear regression and a regional averaging method in predicting R0 using 4-fold cross validation. R0 was <5, 5-10 and >10 for 81, 14 and 3 settings respectively, with no apparent regional differences and in the limited available data, it was usually lower for rural than urban areas. R0 was most correlated with educational attainment, and household indicators for the Pearson and Spearman correlation coefficients respectively and with poverty-related indicators followed by the crude death rate considering the Maximum Information Coefficient, although the correlation for each was relatively weak (Pearson correlation coefficient: 0.4, 95%CI: (0.24,0.48) for educational attainment). A random forest did not perform better in predicting R0 than simple linear regression, depending on the subsets of training indicators and studies, and neither out-performed a regional averaging approach. R0 for rubella is typically low and using indicators to estimate its value is not straightforward. A regional averaging approach may provide as reliable an estimate of R0 for settings lacking seroprevalence data as one based on indicators. The findings may be relevant for other infections and studies estimating the disease burden and the impact of interventions for settings lacking seroprevalence data.

Entities:  

Mesh:

Year:  2022        PMID: 35239641      PMCID: PMC8893344          DOI: 10.1371/journal.pcbi.1008858

Source DB:  PubMed          Journal:  PLoS Comput Biol        ISSN: 1553-734X            Impact factor:   4.475


Introduction

The impact of an intervention against an infection greatly depends on the basic reproduction number (R) of the associated pathogen, defined as the average number of secondary infectious people resulting from the introduction of a typical infectious person into a totally susceptible population[1]. For endemic vaccine-preventable infections, it is often estimated for a given setting from country-specific seroprevalence data collected before vaccination has been introduced[2], but it is unclear if estimates from one setting can be extrapolated to other countries. An understanding of how the basic reproduction number differs between settings is important for studies which try to either estimate the burden of an infectious disease or to predict the impact of interventions for settings without pre-existing seroprevalence data. For settings without pre-vaccination seroprevalence data, studies sometimes base the pre-vaccination epidemiology and therefore R on the regional average, calculated from the available data from all settings in the same geographical region[3-5]. It is also plausible that individual or combinations of socio-economic indicators could give insight into R for a given setting. For example, the number of people that each person contacts might be expected to be correlated with factors such as the amount of crowding, which, in turn, is often correlated with other factors, such as poverty. To our knowledge, the extent to which the regional average or particular indicators might predict what R might be for a given setting has not been studied for any infection. The only related study is that of Santermans et al[6], which studied how differences in R for varicella zoster for 9 European countries might be explained by differences in demographic, socio-economic and spatio-temporal indicators. The study found positive associations between R and factors such as infant vaccination coverage for different vaccines and childcare attendance and negative associations with wealth inequality and poverty. As the study considered a small number of European countries and one infection, it is unclear whether the findings are generalisable either to other continents or to other infections. In this study, we estimate the basic reproduction number for rubella for 98 settings from around the world and correlate its value against 66 demographic, economic, education, health and housing indicators. In addition, we compare the performance of simple linear regression and random forest approaches in predicting R from these 66 indicators against R estimated using a regional averaging approach.

Results

Estimates of the basic reproduction number

Fig 1 summarizes estimates of the basic reproduction number for each study. The basic reproduction number was less than 5 for over half of the settings, with the point estimate being below 5 for 81 settings and the upper 95% confidence limit being below 5 for 62 settings respectively. The point estimate of the basic reproduction number was in the range 5–10 for 14 settings and exceeded 10 for just 3 settings, namely rural Chile, East Germany, and the Czech Republic before 1967, although the confidence intervals were very wide for each setting.
Fig 1

Estimates of the basic reproduction number for each of the studies.

The blue circles reflect the values calculated from study-specific seroprevalence data (study-specific point estimates), the red circles reflect the default R estimates and the vertical markers reflect R estimated from the regional point estimate of the force of infection. The red and blue bars reflect the 95% ranges of the study-specific and default estimates based on bootstrapping.

Estimates of the basic reproduction number for each of the studies.

The blue circles reflect the values calculated from study-specific seroprevalence data (study-specific point estimates), the red circles reflect the default R estimates and the vertical markers reflect R estimated from the regional point estimate of the force of infection. The red and blue bars reflect the 95% ranges of the study-specific and default estimates based on bootstrapping. For most of the settings for which data were available for both urban and rural areas from the same year, the basic reproduction number was higher for urban than for rural areas, although the 95% confidence intervals sometimes overlapped. For urban and rural Peru before 1967, for example, it was 3.7 (95% CI: 3.3–4.4) and 2.2 (95% CI: 1.9–2.4) for respectively, and 6.5 (95% CI: 5.4–7.8) and 2.3 (95% CI 2.1–4.0) for urban and rural Uruguay before 1967 respectively. Exceptions to this pattern included Chile from before 1967 for which the basic reproduction number was 3.9 (95% CI: 3.5–5.2) and 16.5 (95% CI: 12.8–25.2) for urban and rural areas respectively. In general, the value for the basic reproduction number that would have been estimated for each setting using the default method was approximately 2.3–3.3 for countries in the African, Eastern Mediterranean, Western Pacific and South East Asian regions, and around 2.7–3.8 and 3.7–6.7 for countries in the Americas and Europe respectively. The 95% range of the estimates based on the default approach, however, was very wide, with the upper limit of the 95% range reaching over 30 for several studies. The corresponding mean square error, calculated over all the studies, of R for each setting based on the default approach compared to R calculated using study-specific seroprevalence data was approximately 7, meaning that on average, estimates of R of 4, for example, could be predicted to be as low as 1.5 or as high as 6.5. However, the error varied between regions (see S3 Table), with the lowest MSE of around 1 being predicted for the African and Western Pacific regions, intermediate values of 5 predicted for the Eastern Mediterranean and South East Asian regions, and high values of 9 and 23 predicted for the Americas and Europe. The range of the MSE was also very wide and differed between regions, with a 95% range of 1–16 and 13–695 from Africa and Europe respectively (S3 Table). The same general patterns in the basic reproduction number were estimated for the alternative assumptions about contact between children and adults (S1 Fig), with R being below 10 for most settings. For pessimistic assumptions about contact between children and adults, very high values for R of above 20 were estimated for two settings, namely rural Chile and the Czech Republic before 1967. For both assumptions about contact, estimates of R using the regional point estimate of the force of infection were usually smaller and led to a larger mean-square error than the default R (Figs 1 and S1 and S3 Table).

Correlation between R calculated using study-specific seroprevalence data and indicators

In general, the indicators with which R, as estimated using study-specific seroprevalence data, was most correlated depended on the correlation coefficient and, when the correlation was calculated considering all 98 seroprevalence datasets, the correlation was generally weak (Tables 1 and S4, and Fig 2). For the Pearson correlation coefficient, the greatest correlation was obtained between the basic reproduction number and the educational attainment among people aged 25+ years (0.4 95% CI: 0.24, 0.48). For the Spearman correlation coefficient, three housing-related indicators had the highest correlation with the basic reproduction number, with the number of physicians per 1000 population having the next greatest correlation (0.32, 95% CI: 0.22,0.37). Considering the MIC, the basic reproduction number had the greatest correlation with two poverty-related indicators, namely Poverty gap at $1.90 a day (0.37, 95% CI: 0.21,0.34) and Poverty headcount ratio at $5.50 a day (0.36, 95% CI: 0.22,0.37).
Table 1

Summary of the top ten correlation coefficients and MIC for the association between the basic reproduction number and the indicators, obtained considering estimates of the basic reproduction number for all countries.

The columns labelled “CC” hold the coefficient, with the 95% range obtained by bootstrapping; the column labelled “N” holds the number of data points used to calculate the coefficient.

RankPearsonSpearmanMIC
IndicatorCCNIndicatorCCNIndicatorCCN
1Educational attainment, at least completed upper secondary, population 25+, total (%) (cumulative)0.4(0.24, 0.48)88Number of households 5 persons—Proportion over All households-0.45(-0.5, -0.36)56Poverty gap at $1.90 a day (2011 PPP) (%)0.37(0.21, 0.34)92
2Educational attainment, at least completed upper secondary, population 25+, female (%) (cumulative)0.4(0.25, 0.47)88Number of households 5 persons—Per capita-0.41(-0.46, -0.32)56Poverty headcount ratio at $5.50 a day (2011 PPP) (% of population)0.36(0.22, 0.37)92
3Educational attainment, at least completed upper secondary, population 25+, male (%) (cumulative)0.39(0.22, 0.47)88Number of households 6 persons and over—Per capita-0.33(-0.41, -0.22)53Crude death rate per 1000 population0.35(0.2, 0.37)98
4Number of households 5 persons—Proportion over All households-0.34(-0.42, -0.19)56Physicians (per 1,000 people)0.32(0.22, 0.37)98Life expectancy at birth (both sexes)0.34(0.2, 0.33)98
5Physicians (per 1,000 people)0.33(0.16, 0.41)98Prevalence of underweight, weight for age (% of children under 5)-0.31(-0.34, -0.21)94Poverty gap at $3.20 a day (2011 PPP) (% of population)0.34(0.22, 0.36)92
6Proportion of the population aged 65+0.32(0.2, 0.51)98Immunization, measles (% of children ages 12–23 months)0.3(0.22, 0.36)98Poverty headcount ratio at $1.90 a day (2011 PPP) (% of population)0.34(0.22, 0.36)92
7Number of households 5 persons—Per capita-0.28(-0.35, -0.1)56Number of households 6 persons and over—Proportion over All households-0.29(-0.37, -0.18)53Poverty gap at $5.50 a day (2011 PPP) (% of population)0.33(0.23, 0.37)92
8Proportion of the population aged 0–14-0.27(-0.39, -0.14)98Health expenditure, total (% of GDP)0.28(0.21, 0.35)98Poverty headcount ratio at $3.20 a day (2011 PPP) (% of population)0.33(0.22, 0.37)92
9Population living in slums (% of urban population)-0.27(-0.32, -0.13)75Urban population (% of total)0.28(0.2, 0.34)98Proportion of the population aged 0–40.32(0.27, 0.38)98
10Number of households 6 persons and over—Proportion over All households-0.26(-0.34, -0.14)53Educational attainment, at least completed upper secondary, population 25+, total (%) (cumulative)0.27(0.2, 0.35)88Physicians (per 1,000 people)0.31(0.25, 0.37)98
Fig 2

Indicators in each category with the largest Spearman correlation with R (correlation value and CI in red font).

The Pearson correlation regression line is plotted in red.

Indicators in each category with the largest Spearman correlation with R (correlation value and CI in red font).

The Pearson correlation regression line is plotted in red.

Summary of the top ten correlation coefficients and MIC for the association between the basic reproduction number and the indicators, obtained considering estimates of the basic reproduction number for all countries.

The columns labelled “CC” hold the coefficient, with the 95% range obtained by bootstrapping; the column labelled “N” holds the number of data points used to calculate the coefficient. When the simple correlation analyses were repeated by region, economic indicators were the most correlated with the basic reproduction number in Africa (Table 2). For the Americas, health-related indicators, for example, the percentage of children who were aged under 6 months who were breast-feeding, the percentage of the population that was urban and several poverty-related indicators were most correlated with the basic reproduction number (Table 2). For several indicators, the size of the correlation with R was larger when they were calculated on a regional, compared to a global basis (Tables 1 and 2 and Fig 2). For the remaining regions, the small number of studies complicated the interpretation of the findings from linear regression (findings not shown).
Table 2

Summary of the top five correlation coefficients and MIC for the association between the basic reproduction number and the indicators, obtained considering estimates of the basic reproduction number just for countries in the African region or the Americas.

The columns labelled “CC” hold the coefficient, with the 95% range obtained by bootstrapping; the column labelled “N” holds the number of data points used to calculate the coefficient.

RankPearsonSpearmanMIC
IndicatorCCNIndicatorCCNIndicatorCCN
African region
1Unemployment, total (% of total labor force) (national estimate)0.35(0.09, 0.48)24Poverty headcount ratio at national poverty lines (% of population)-0.4(-0.46, -0.18)24Number of households 3 persons—Proportion over All households1.0(0.19, 1.0)6
2Poverty headcount ratio at national poverty lines (% of population)-0.33(-0.44, -0.08)24Income share held by highest 10%0.38(0.17, 0.5)24Number of households 4 persons—Proportion over All households1.0(0.19, 1.0)6
3Income share held by highest 10%0.31(0.09, 0.54)24Health expenditure, total (% of GDP)0.38(0.19, 0.52)24Immunization, measles (% of children ages 12–23 months)0.55(0.23, 0.65)24
4Poverty gap at $1.90 a day (2011 PPP) (%)-0.3(-0.36, -0.08)24Income share held by highest 20%0.33(0.15, 0.51)24Unemployment, total (% of total labor force) (national estimate)0.52(0.28, 0.59)24
5Poverty gap at $3.20 a day (2011 PPP) (% of population)-0.29(-0.34, -0.06)24Educational attainment, at least completed upper secondary, population 25+, male (%) (cumulative)-0.32(-0.45, -0.06)19Physicians (per 1,000 people)0.49(0.27, 0.57)24
Americas
1Exclusive breastfeeding (% of children under 6 months)0.45(0.3, 0.52)23Urban population (% of total)0.61(0.53, 0.69)26Immunization, DPT (% of children ages 12–23 months)0.58(0.36, 0.63)26
2Immunization, DPT (% of children ages 12–23 months)0.45(0.19, 0.51)26Exclusive breastfeeding (% of children under 6 months)0.59(0.52, 0.65)23Exclusive breastfeeding (% of children under 6 months)0.55(0.42, 0.6)23
3Unemployment, total (% of total labor force) (modeled ILO estimate)-0.42(-0.5,-0.28)26Prevalence of underweight, weight for age (% of children under 5)-0.57(-0.67, -0.47)26Poverty gap at $3.20 a day (2011 PPP) (% of population)0.53 (0.41, 0.53)26
4Educational attainment, at least completed upper secondary, population 25+, total (%) (cumulative)0.42(0.21, 0.51)24Unemployment, total (% of total labor force) (modeled ILO estimate)-0.57(-0.62, -0.48)26Poverty headcount ratio at $3.20 a day (2011 PPP) (% of population)0.53(0.41, 0.53)26
5Urban population (% of total)0.42(0.28, 0.51)26Immunization, DPT (% of children ages 12–23 months)0.49(0.39, 0.59)26Adjusted net enrollment rate, primary (% of primary school age children)0.48(0.32, 0.57)25

Summary of the top five correlation coefficients and MIC for the association between the basic reproduction number and the indicators, obtained considering estimates of the basic reproduction number just for countries in the African region or the Americas.

The columns labelled “CC” hold the coefficient, with the 95% range obtained by bootstrapping; the column labelled “N” holds the number of data points used to calculate the coefficient.

Comparison between R predicted using simple linear regression and random forest regression

Figs 3 and 4 compare the performance of simple linear regression, random forest and the default method for calculating R, as quantified by the MSE over the 4-fold cross-validation experiments, of the value predicted by these methods compared to that calculated using study-specific seroprevalence data. In general, the prediction error associated with using the default approach for calculating R is comparable and in many instances slightly lower than the corresponding linear regression and random forest error, although the range of the error sometimes overlapped.
Fig 3

Mean value (blue and orange dots) and minimum-maximum value range (blue and orange line) of the MSE of the predicted R over the 10 repetitions of the 4-fold cross-validation experiment using the non-imputed dataset.

The blue color corresponds to the 66 linear regression results (one for each indicator) and the orange color to the 5 random forest results (one for each subset of indicators). The green dots indicate the average MSE for R as calculated using the default approach compared to that calculated using study-specific seroprevalence data.

Fig 4

Mean value (blue and orange dots) and minimum-maximum value range (blue and orange line) of the MSE of the predicted R over the 10 repetitions of the 4-fold cross-validation experiment using the imputed dataset.

The blue color corresponds to the 66 linear regression results (one for each indicator) and the orange color to the 5 random forest results (one for each subset of indicators). The green dots indicate the average MSE for R as calculated using the default approach compared to that calculated using study-specific seroprevalence data.

Mean value (blue and orange dots) and minimum-maximum value range (blue and orange line) of the MSE of the predicted R over the 10 repetitions of the 4-fold cross-validation experiment using the non-imputed dataset.

The blue color corresponds to the 66 linear regression results (one for each indicator) and the orange color to the 5 random forest results (one for each subset of indicators). The green dots indicate the average MSE for R as calculated using the default approach compared to that calculated using study-specific seroprevalence data.

Mean value (blue and orange dots) and minimum-maximum value range (blue and orange line) of the MSE of the predicted R over the 10 repetitions of the 4-fold cross-validation experiment using the imputed dataset.

The blue color corresponds to the 66 linear regression results (one for each indicator) and the orange color to the 5 random forest results (one for each subset of indicators). The green dots indicate the average MSE for R as calculated using the default approach compared to that calculated using study-specific seroprevalence data. When the missing indicators were not imputed (Fig 3), the error associated with predictions based on linear regression differed between indicators. For most indicators, the MSE values range from approximately 7.5 to 9. The smallest error (MSE = 2.73) was associated with a poverty-related indicator (“Poverty gap at national poverty lines (%)”). The errors associated with four indicators related to schooling, undernourishment, breastfeeding and living in slums were also considerably lower than most, with a MSE of 5.35–5.5. The errors associated with the indicators for the number of doctors’ consultations, number of people per room and all the indicators related to number of households with various numbers of persons are higher than all the rest, ranging in values of 11 and upwards. The min-max range of the error is fairly narrow and the minimum MSE in badly performing indicators is generally higher than the maximum MSE of better performing indicators. In the random forest non-imputed results, the lowest average MSE equals 6.88 for the random forest trained and tested using the 63 indicators that have up to 70 missing values and the highest average MSE is equal to 13.8 for the random forest trained and tested using the 50 indicators with up to 40 missing values. In contrast, in the imputed data results (plotted in Fig 4) the MSE for all linear regression, random forest and default methods is similar at around the value of 8. As shown in Figs A and B in S2 Text much of the variation in the performance of the indicators in the simple linear regression and random forest analyses disappears once the two studies which have much higher values for R than the other studies, namely, the Czech Republic and Chile (rural) (R = 19.97 and R = 16.53 respectively) are omitted from the experiments. The individual prediction error of those two studies drives the overall MSE of the 4-fold cross validation experiment upwards. As can be seen in Table A in S2 Text, the best performing indicator is the only one having a missing value for both those studies. In addition, the difference in the MSE between the default approach for predicted R and simple linear regression and random forest methods reduces once these two studies are omitted from the experiments. Further details of the effect of the presence or absence of these two studies in the folds are provided in S2 Text.

Random forest parameter tuning

In general, optimising the set of parameters used for the random forest algorithm for predicting R led to a small reduction in average MSE, from 9.87 for the default set of parameters to 8.86 in the optimised case (these values correspond to the average of the MSE for the 10 repetitions plotted in Fig 5). In addition, the latter of those values was still higher than that associated with the default approach used for calculating R, for which the MSE was 6.97 (as plotted in Fig 3). As shown in Fig 5, the optimal value for the MSE obtained with the nested-optimisation method for the random forest algorithm did not vary considerably across the 10 repetitions, although the minimum to maximum range of the MSE across the 96 different sets of parameters varied. A more detailed exposition of the nested cross-validation optimisation results is given in S3 Text. We note that the minimum end of the performance range (blue bars in Fig 5) is lower than the MSE error of the nested optimisation results in some of the 10 experiment repetitions. This is in accordance with previous theoretical and empirical comparison results between nested and non-nested optimisation methods[7].
Fig 5

Performance (as quantified by the MSE in each of the 10 splits to folds) of the default set of parameter values used in the random forest (yellow x) compared to the optimal parameters for the random forest identified through nested optimisation (blue circle).

Blue lines are the MSE minimum to maximum range for the 96 points of the parameters grid (hence this minimum is the optimal result of the non-nested optimization method).

Performance (as quantified by the MSE in each of the 10 splits to folds) of the default set of parameter values used in the random forest (yellow x) compared to the optimal parameters for the random forest identified through nested optimisation (blue circle).

Blue lines are the MSE minimum to maximum range for the 96 points of the parameters grid (hence this minimum is the optimal result of the non-nested optimization method).

Discussion

Our analyses found that the basic reproduction number for rubella is typically low and under 5 for many settings, which is consistent with the view that rubella is less infectious than other childhood immunizing infections, such as measles and mumps. We also found that in many cases, the basic reproduction number was lower for rural than areas but that the correlation between R and indicators was not straightforward. In general, the performance of a regional averaging approach to estimate the basic reproduction number was often better than that using linear regression and random forest approaches. This is the first study to attempt to correlate the basic reproduction number for rubella in different time periods globally against different indicators. The only other related study to date, by Santermans et al considered varicella zoster and considered nine European settings[6]. The study found positive associations between R and factors such as infant vaccination coverage for different vaccines and childcare attendance and negative associations with wealth inequality and poverty. As shown in our analyses, the correlation between indicators and R for rubella is not straightforward, with the indicator with which R had the greatest correlation depending both on the correlation statistic used and the region. In addition, the size of the correlation depended on the region considered, and was weaker when all study settings were considered than when the correlation was conducted on a regional basis. The size of the correlation in the latter case was comparable to that seen in Santermans et al[6]. The correlation between the basic reproduction number and several indicators was in the opposite direction from that expected. For example, in the simple linear regression analyses considering all countries, the basic reproduction number went down as the proportion of households which comprised 5 people increased and it increased as the number of physicians per 1000 population increased. This may have been due to other unknown factors confounding the observed relationship between R and crowding indicators. It is also plausible that some outliers may have affected our results, but the extent to which this was the case is unclear. For example, considering Singapore, the estimated R was low but the population density was much higher than in other settings. Our finding that the basic reproduction number was typically higher in urban than in rural areas (Fig 1) is consistent with the fact that the people in urban areas are more likely to contact others than people in rural areas. One of the exceptions to this was Chile from before 1967, for which R in rural areas was several times greater than that in rural areas. An unexpectedly high estimate in rural areas might occur if, for example, the seroprevalence data had been collected shortly after an epidemic had occurred, resulting in unusually high levels of seroprevalence. Small differences in the R between urban and rural areas may also occur if rural residents often visit urban areas, because of work or other reasons. To our knowledge, our analyses are the first to assess the performances of regional averaging (default), simple linear regression and random forest approaches in predicting R. We found that the default regional averaging approach employed in this work, whereby we calculated a set of R estimates based on each force of infection estimate from the regional collection of bootstraps and then took the median of that R estimate, performed better than the “regional point estimate of R”. The under-performance of the latter is unsurprising, given its method of computation using the regional “point estimates” of the force of infection. These were defined in previous analyses[5,8-10], in which the main outcome of interest of interest was the incidence of Congenital Rubella Syndrome (CRS), as the force of infection estimates which led to the median incidence of CRS per 100,000 live births in the absence of vaccination. Since the CRS incidence depends both on the proportion of women who reach child-bearing age still susceptible and the force of infection among adult women, the combination of force of infection estimates among children and adults, out of 1000 bootstrap-derived values, which results in the median CRS incidence is unlikely to result in the median estimate of the basic reproduction number. Although the two machine learning approaches are technically more sophisticated than the default approach it is interesting that, when all studies were considered, they were often out-performed by the default approach, based on regional averaging. However, we also note that the two machine learning approaches were sensitive to inclusion of two datasets with particularly high R values and once they were excluded, the performance of the three approaches was similar. These findings suggest that for studies of the disease burden or the impact of interventions against rubella, indictor-based approaches are unlikely to improve upon using a regional averaging approach for settings lacking seroprevalence data. There are several limitations of our analyses, which could have contributed to the indicator-based approaches under-performing when predicting R. For example, for simplicity, the value for the indicator used when correlating the basic reproduction number was taken from the year in which the study was conducted. In practice, the seroprevalence in the study year results from the exposure to rubella infection over the lifetime of the study population. In addition, when calculating the basic reproduction number, we just used one assumption for the contact between adults and children. Whilst that assumption is based on estimates from social contact studies from the past decade, it is plausible that for some of the settings in our study, such as those from many years ago, the amount of contact between people would have been different compared to that in recent years. In addition, due to the limitations of the seroprevalence data, which typically comprised convenience samples covering three or four age groups, we used a crude age stratification when calculating R, just considering two broad age groups. Our estimates of R using the default approach involved taking the median of R values as calculated using bootstrap-derived estimates of the force of infection from all studies in the same World Health Organization region. The performance of this approach may have been overestimated given that bootstrap-derived estimates used to calculate the median would have included those from each of the studies being used to evaluate the performance. In practice, the overestimate is likely to have been relatively small, given that at least 10 studies would have contributed to the bootstrap samples in each region. In order to corroborate this, we included in our analyses a calculation of the default R estimates for all the studies of one region (EMRO region) in which the force of infection bootstraps from each study are excluded from the collection of bootstraps used to derive the default R estimate of that particular study (see S4 Text). As is seen in Fig A in S4 Text, the impact of this exclusion is very small both in the cases where the default R estimate is close and further way from study-specific R estimate value. In addition to that, we note that in analyses of the global burden of CRS, which included the datasets using the ones in this study, excluding a single seroprevalence dataset had little impact on estimates[4]. In a related manner, the comparison of the performance of the default approach to those of the Machine Learning methods using a cross-validation methodology is possibly giving an advantage to the former. This is because the default R estimates for each study are calculated once, prior to the cross-validation split to folds, and they make use of information in studies from the same region regardless of whether these fall in the test or training set in each particular split to folds. A more accurate version of that comparison would be to modify the default approach so that for each different indicator and each different split to cross-validation folds, a different default R estimate is calculated for each study by excluding from the regional collection of force of infection bootstraps the studies that are in the same test fold (this exclusion would be dependent on the different indicators because of the missing values issue). However, given the minimal impact of excluding force of infection estimates of one study at a time (see discussion above and in S4 Text), and the increased complexity of the aforementioned modification, we did not pursue it. The method we used to calculate the study-specific R estimates had a few limitations: Our analyses did not take into account the genetic diversity of the Rubella virus[11] which may be a source of variability of the basic reproduction number over different settings, that could not have been predicted by the indicator that we used as inputs to the machine learning methods we examined. Another limitation is that, while the seroprevalence data used to calculate the force of infection estimates used in the R typically spanned three or more age groups we used two age groups to describe the age-mixing patterns. The influence of more complex mixing patterns has been described elsewhere[12]. Finally, the possible effect of different assays in the seroprevalence data that we used for the determination of the study-specific R estimates[13] may have introduced a source of variability not predictable by our machine learning methods. Several further refinements of both the default and indicator-based approaches for predicting R are potentially feasible, although the effect on their performance is unclear. First, the default approach could be refined to use alternative methods for grouping the countries, such as those developed for Global Burden of Disease analyses, which are based on both epidemiological aspects and geographical location. However, application of this method for grouping countries is complicated by the fact that some groupings include only a few countries each of which may have no pre-vaccination seroprevalence data. Second, the parameter optimisation experiment that we ran on the random forest regression algorithm was limited to only 6 parameters and to 2 or 3 values for each of those parameters. A search over a more extensive parameter grid (even though much more computationally intensive) could result in a reduction of the expected prediction error of the random forest algorithm. However, the small scale parameter optimisation conducted here suggests that the prediction results obtained with the random forest prediction method are close to the optimum and that not much further improvement can be expected by a finer or wider parameter grid search. Third, the effect of different methods for imputing missing values could be explored. As presented in the results, treating missing values with median value imputation has a significant effect on the prediction performance (Figs 3 and 4) and the effect of different methods of imputation on the prediction performance could be investigated. Fourth, the prediction methods that we considered are unable to incorporate the uncertainty inherent in our data, which is present both in the indicator data that we use as input and in the estimated R values that we use as the ground truth. Considering input data, the uncertainty can be quantified as the number of years of difference between the seroprevalence and the indicator data and this could also be modelled to fold in the case of missing data. In the case of the output the uncertainty can be quantified by the bootstrap-derived data distribution for different values of the contact parameter k. A Gaussian process methodology could be employed to incorporate those measures of uncertainty in the regression problem. The main advantage of such a method would be that it can provide a measure of uncertainty of the predicted results. For the prediction to be improved, it would be important for these measures of uncertainty to overcome one of the limitations of each of the prediction methods, namely that they are unable to correctly predict the very high values of R for some settings (particularly for those in which R>10). In conclusion, our analyses confirm the view that the basic reproduction number for rubella is low and that an approach based on regional averaging for settings lacking prevaccination seroprevalence is likely to perform as well as, if not better, than indicator-based approaches involving simple linear regression or random forest regression algorithms. It is unclear whether refinement of these methods could lead to an improvement in their performance. Whilst our analyses have focused on rubella, these findings are likely to be relevant for studies of the disease burden or the impact of interventions for other infections.

Methods

Data sources

Seroprevalence data

The analyses use force of infection estimates (defined as the rate at which susceptible people are infected) for 98 rubella seroprevalence datasets that were identified in a systematic review and had been used previously in estimates of the global burden of Congenital Rubella Syndrome (CRS) and the impact of measles-rubella vaccination[4,5]. The seroprevalence datasets are listed in Table A in S1 Table and were collected from 24, 26, 13, 13, 13 and 9 settings from the African, the Americas, Eastern Mediterranean, Europe, South East Asia and Western Pacific World Health Organization regions respectively before the introduction of rubella vaccination.

Demography

The age distribution considering single year age bands of each of the countries used in calculating the basic reproduction number was based on UN population data[14] for the year in which the study was conducted or the publication year, if the study year was unavailable. For instances in which the number of people aged 80 years and over is not stratified into single year age bands, we calculated this stratification using UN population data on the number of survivors by age.

Indicators

For each of the 98 seroprevalence studies, we extracted 66 demographic, economic, education, health and housing indicators for the year in which the study was conducted, where possible. The indicators expand on 39 indicators used by Santermans et al in their study of R for varicella-zoster[6] and are summarized in Table A in S2 Table. For studies for which the study year was not known, the indicator was extracted for the year in which the data were published. If the value of the indicator for the actual study year was not available, the indicator was extracted for the year closest to the study year. Educational and employment were extracted from World Bank sources[15]. Economic and health-related indicators were extracted mainly from World Bank sources[15] but also from United Nations, International Monetary Fund, Organisation for Economic Co-operation and Development, World Health Organization and other sources[16-20]. Demographic data were extracted from United Nations population data and World Bank sources[14,15]. Housing data were extracted from World Bank and United Nations Statistics Division sources[15,21,22]. Table A in S2 Table lists the data source for each indicator and it summarizes the completeness of data on the indicators for the studies and the number of studies for which values for the indicators in question were available within 5 years, 5–10 and >10 years of the study year. The number of indicator values available for each study are provided in Table B in S2 Table. Values for the indicators were available for almost all of the studies, except for many of the housing-related indicators, which were not available for approximately 45% of the study settings. Values for all but one of the demographic indicators were available within 5 years of the study being conducted for all settings, and values for some of the economic indicators were only available more than 10 years after the study was conducted for 50% of the studies.

Estimating the basic reproduction number

For each of the 98 seroprevalence studies, point estimates of the force of infection among those aged <13 and ≥13 years, as published elsewhere[4,5], were used to calculate the basic reproduction number, as the dominant eigenvalue of the Next Generation Matrix[1]. We refer to the estimates of R derived from the point estimates of the force of infection from each study as “study-specific point estimates of R”. Additionally, 1000 bootstrap-derived samples for the force of infection (calculated as described in[4]) were used to derive 1000 bootstrap-derived estimates of the study-specific R and, subsequently, the 95% range of the study-specific R for each study. The forces of infection for each study are summarised in Fig 6. The matrix of Who Acquires Infection from Whom was assumed to have the following structure:
Fig 6

Summary of the force of infection estimates used to calculate the basic reproduction number for each study in each region.

The bars reflect the 95% ranges based on bootstrapping.

Summary of the force of infection estimates used to calculate the basic reproduction number for each study in each region.

The bars reflect the 95% ranges based on bootstrapping. The effective contact rate differs between <13 and ≥13 year olds, with its relative size between children and adults, compared to that between adults (k), assumed to equal 0.7, based on contact survey data[23]. In sensitivity analyses, we explored the effect of using pessimistic assumptions about mixing between younger and older people on estimates of the basic reproduction number, using a value of 0.3 for k. The age distribution of the population used in calculating the basic reproduction number was based on UN population data[14] for the year in which the study was conducted or the publication year, if the study year was unavailable. Previous studies have approximated the epidemiology of rubella in settings lacking seroprevalence by using a regionally averaging approach, using 1000 bootstrap-derived samples of the forces of infection compiled from all the studies from the same World Health Organization region[4,5]. To evaluate this approach, for each of the 98 studies we calculated what the median R and its 95% range would have been estimated to be from these 1000 bootstrap-derived force of infection samples drawn from all the countries of the corresponding region. The bootstraps drawn from each study to create the each of those regional collection of 1000 force of infection bootstrap values is described in detail in references[4,5]. We refer to this median R as the “default R” throughout the text. These default estimates of R for the 98 studies are the same throughout our analyses including the cross-validation results we present. The error associated with the default R and that calculated using the study-specific point estimate of R was quantified as the mean square error (MSE) over the 98 studies. The 95% range of the MSE was calculated using 1000 bootstrap-derived estimates of the MSE computed from the corresponding R estimates that were computed from 1000 bootstrap-derived force of infection. In sensitivity analyses, we also calculate what R for a given setting would have been estimated to be using a point estimate of the age-dependent force for the region, which has previously been used to obtain an approximate estimate of the average CRS incidence for countries lacking seroprevalence data[5,8]. This regional point estimate of the force of infection was taken as the value which led to the median burden of CRS out of the 1000 bootstraps in the absence of vaccination (see supplement in ref[5]). We refer to the estimate of the basic reproduction number obtained using the regional point estimate of the force of infection as the “regional point estimate of R”

Measures of relation between indicators and R estimates

The Pearson and Spearman correlation coefficients and the Maximum Information Coefficient (MIC) were calculated for the correlation between the basic reproduction number for each study and each of the 66 demographic, economic, education, health and housing indicators extracted for each study. The MIC has been used elsewhere in similar analyses for varicella zoster[6] and quantifies the correlation between two statistics when they may be non-linearly related[24]. The 95% range of the correlation coefficients were calculated using the 1000 bootstrap-derived values for the basic reproduction number. In the first instance, the correlation coefficients were calculated considering the basic reproduction numbers for all countries. In sensitivity analyses, the regression was repeated on a regional basis. In the results we present which rank the level of relation between the different indicators and R from highest to lowest, we rank lowest all the indicators for which the 95% CI overlaps zero (i.e. the 2.5% and 97.5% percentiles of the bootstrap values are of opposite sign).

Out-of-sample prediction

Overview

We explored whether the basic reproduction number for a given setting could be predicted using either simple linear regression on any of the 66 indicators or a Random Forest regression algorithm trained on different subsets of the indicators. We used 4-fold cross-validation[25] to score both approaches and explored the effect of imputation. Further details of the cross-validation and imputation are provided below.

Four-fold cross-validation

The 4-fold cross-validation process was similar for both the simple linear regression and random forest algorithms. A detailed description of the process (“experiment”) is provided in S1 Text and we summarize it briefly here. First, the values for R that had been estimated using study-specific seroprevalence data for the 98 studies, along with their indicators, were split into 4 partitions (“folds”). Next, for the case of simple linear regression and for a given indicator, the values in three of the folds were used as the training dataset to establish a relationship between the point estimate of R that had been estimated using seroprevalence data, and that indicator. The values in the fourth fold were then used as the test dataset, whereby the value of the indicator for each study was used to predict what R might be expected to be in that study using the linear regression model that had been estimated using the training dataset. We then calculated the mean square error (MSE) for the test dataset in these predicted R values compared to those estimated using seroprevalence data. These steps were repeated using each fold in turn as the test dataset and the remaining three as the train dataset, and we computed the average MSE for the four folds. These steps were repeated for each of the indicators in turn. The four-fold cross- validation was repeated using 10 randomly generated splits of the 98 R estimates to 4 folds, resulting in one value of the average MSE for each split and in the final stage of this process, we computed the average, minimum and maximum of those 10 average MSE values of the predicted R compared to that calculated from study-specific seroprevalence data. We discarded splits to folds that had insufficient numbers of indicator values in each fold to train and test the linear regression algorithms (see further details below). We used the steps and the same splits to folds as described above for estimating the performance of the random forest regression algorithm, except that subsets of indicators (rather than a single indicator at a time) are used train and test the random forest regression algorithm, as described below.

Dealing with missing values in the indicator data

Most of the 98 studies lacked values for some of the indicators (see Table B in S2 Table) and we explored the effect of both not imputing the missing values and of imputing them as follows. A. Non-imputation approach. Not imputing the missing values has implications in the 4-fold cross-validation experiments. Considering the situation for simple linear regression, we note that the situation may arise where for a given indicator and a given split of the 98 studies to 4 folds, the training dataset has valid values for only one or none of the studies in it. In such cases it is not possible to fit a regression line. To address this, we generated a series of splits to folds and we kept the first 10 splits for which each indicator had at least one study with a valid value in each of the 4 folds (see S1 Text for further details). This guarantees that for all folds and all indicators there are at least 3 studies with valid values in the training dataset which is sufficient for fitting a regression line and that there is at least one study with a valid value in the test dataset. With this arrangement, a study that has a valid value for a given indicator is used exactly once for testing and three times for training the linear regression model corresponding to that indicator in each of the 10 repetitions of the 4-fold cross-validation experiment. For consistency, we used the same 10 splits to folds for the random forest 4-fold cross-validation experiments. Since random forest regression algorithms use all available indicators to train and predict the outcome of interest, rather than just one, as is the case for simple linear regression, the treatment of the missing data was different from that used for simple linear regression. Instead, we considered the following 5 different subsets of indicators and we ran a separate cross-validation experiment for each subset: The 25 indicators that have no missing values at all, i.e. all 98 studies have values for those indicators. The 43 indicators that have up to 10 missing values. There are 69 studies which have values for each of those 43 indicators. The 46 indicators that have up to 20 missing values. There are 52 studies which have values for each of those 46 indicators. The 50 indicators that have up to 40 missing values. There are 32 studies which have values for each of those 50 indicators. The 63 indicators that have up to 70 missing values. There are 20 studies which have values for each of those 63 indicators. We list the number of studies in the training and testing subsets for all 10 splits to 4 folds for both the linear regression and the random forest models in Table A in S1 Text. B. Imputation approach. In the second approach we use imputation and we replaced all missing values (if any) of each indicator with the median value of the existing values among the 98 studies for that particular indicator. The imputation is applied once to the whole dataset prior to splitting to folds and all other parts of the cross-validation. With this approach there are no missing values in the imputed dataset and we were able to conduct all training and testing parts of the 4-fold cross-validation experiments as usual. For consistency purposes we used the same choice of 10 4-fold splits of the non-imputed experiments that is described above.

Comparison of the linear regression, random forest R0 predictions against those from the default method

To assess whether the linear regression and random forest approaches performed better in predicting R across the settings than using the default R estimate, we computed the average, minimum and maximum values of MSE over the same test folds of the 10 cross-validation splits described above for the default R estimate compared to R estimated using study-specific seroprevalence data. The resulting MSE values were compared against those calculated considering the R for the linear regression and random forest approaches. As described above, R calculated using the default approach is taken as the median R calculated from 1000 bootstrap-derived samples of the forces of infection compiled from all the studies from the same World Health Organization region.

Random forest optimisation

Unlike the fitting of a linear regression model, which is solely a function of the training dataset, the process of training a random forest prediction model is also a function of the choice of hyper-parameters of the random forest. For the prediction performance comparison results that we described above we used the default parameter values of Python’s scikit-learn toolbox. Further to that, we also present results of a separate small-scale parameter optimisation run. For that parameter optimisation experiment we used the same 10 cross-validation splits to 4 folds described above and the first indicators’ subset listed above, i.e. the one comprising the 25 indicators that have no missing values. We present non-nested optimisation results as well as nested cross-validation optimisation results (see references[7,26]). The former method amounts to simply selecting the parameter setup which achieves the lowest average MSE for a given split to folds. That method effectively violates the out-of-sample principle and can give over-optimistic estimates of the expected prediction performance in previously unseen test data. The latter case amounts to the results obtained when a separate inner-loop cross-validation experiment is run for each cross-validation fold and it is considered a more valid method to estimate the generalisation of the prediction performance in previously unseen test data.

Software

All computations were carried out in a reproducible manner using the corresponding methods and functions of Python’s scikit-learn package[26]. The 10 splits to folds were obtained by use of the ‘sklearn.model_selection.KFold’ class by setting the ‘random_state’ parameter which corresponds to the seed of the pseudorandom generator to increasing integer values starting from 0. The values corresponding to folds that were used in the results are given in Table A in S1 Text. The linear regression and random forest prediction results were obtained with the methods of the ‘sklearn.linear_model.LinearRegression’ and ‘sklearn.ensemble.RandomForestRegressor’ classes. The missing values imputation described in the ‘Imputation approach’ section was implemented using the ‘sklearn.impute.SimpleImputer’ class. For the random forest parameter tuning results we used a 6-parameter grid comprising 96 parameters’ choices (details are given in S3 Text). We note that due to the missing values in the dataset we could not use the higher-level methods of the scikit-learn package (e.g. ‘sklearn.model_selection.cross_val_score’) and that for all the results we had to write code that implements the complete cross-validation methodology.

Estimates of R for pessimistic assumptions about contact.

Fig A: Estimates of the basic reproduction number for rubella, as calculated using seroprevalence data (blue circles) and the default approach (red circles) for each study for pessimistic assumptions about the amount of contact between children and adults k = 0.3). The vertical marks indicate R as calculated using the regional point estimate for the force of infection. (PDF) Click here for additional data file.

Summary of the study settings.

Table A: Summary of the sources of seroprevalence data from each region for the force of infection estimates for each setting, that is used in this analysis. Force of infection estimates from all of these datasets have been previously used in estimating the global burden of Congenital Rubella Syndrome or in the impact of Measles-Rubella vaccination. (PDF) Click here for additional data file.

Completeness of the indicator data.

Table A: Number of studies for which the year in which the indicator was available differed by less than 5 years, 5–10 years or more than 10 years from the year in which the data were collected. For studies with an unspecified year of data collection, the difference was calculated as the difference between the publication year and the year for which the indicator was extracted. The right-most column cites the source from which the data was obtained. Table B: Number of missing indicators for each of the 98 studies. (PDF) Click here for additional data file.

Estimates of the mean square error associated with R estimates.

Table A: Estimates of the mean square error associated with R estimates. Column 2 shows the mean square error of R, as calculated using the default approach against R calculated using study-specific seroprevalence data. Columns 2–7 hold the minimum, median, maximum 2.5th and 95.5th percentiles of the MSE values between the 1000 bootstrap-derived values of R calculated using the regional collection of force of infection bootstrap values and the 1000 bootstrap-derived study-specific R values. The final column holds the MSE between study-specific R0 and the regional point estimate of R. (PDF) Click here for additional data file.

Correlation between R and the indicators.

Table A: Summary of the correlation coefficients and MIC for the association between the basic reproduction number and the indicators. ranked in decreasing order. The columns labelled “CC” hold the coefficient, with the 95% range obtained by bootstrapping; the column labelled “N” holds the number of data points used to calculate the coefficient. (PDF) Click here for additional data file.

Details of 4-fold cross-validation–Dealing with missing indicators in simple linear regression and random forest analyses.

Fig A: Examples of two different splits to folds for the toy-case cross-validation scenario described above. The ‘x’ marks denote missing values. Table A: Number of studies with a valid indicator value in each fold for each of the 10 repetitions of the 4-fold cross validation experiment. The indicators order for the Simple Linear Regression case is the same as in Table A in S2 Table and the order of indicators subsets for the Random Forest case is the same as in the main text. The row labelled ‘Seed’ gives the seed numbers used for each of the 10 split to folds (omitted seed numbers corresponds to splits that did not have at least one study with a valid (non-missing) value in each fold for each one of the 66 indicators. (PDF) Click here for additional data file.

Effect of studies with high R values on the performance of simple linear regression and Random Forest prediction.

Table A: Presence or absence of the two studies with highest value of R from non-imputed linear regression 4-fold cross validation experiments. Table B: Presence or absence of the two studies with highest value of R from non-imputed random forest 4-fold cross validation experiments. Fig A: Similar to Fig 3 of the main text but with the two studies (‘Czech Republic, <1967’ and ‘Chile (rural), 1967–68’) excluded. Mean value (blue and orange dots) and minimum-maximum value range (blue and orange line) of the MSE of the predicted R over ten 4-fold cross validation splits. Fig B: Similar to Fig 4 of the main text but with the two studies (‘Czech Republic, <1967’ and ‘Chile (rural), 1967–68’) excluded. Mean value (blue and orange dots) and minimum-maximum value range (blue and orange line) of the MSE of the predicted R0 over ten 4-fold cross validation splits. (PDF) Click here for additional data file.

Random forest parameter tuning.

Table A: MSE performance and rank of the 96 parameter sets in the double-loop nested cross-validation experiment. The default parameter setup is in bold font. (PDF) Click here for additional data file. Including or excluding the force of infection bootstraps of a study in the calculation of its default R estimate Fig A: Estimates of the basic reproduction number for each of the studies in the EMRO region. The blue and red circles and bars are identical as in Fig 1 of the main text. The black circles and ranges are the default R estimates and 95% CI that we obtain for each study when we exclude bootstrap values from that study from the collection of 1000 force of infection bootstrap used in the calculation. Table A: Comparison of default R values with the values obtained when the regional collection of 1000 force of infections bootstrap values does not include those coming from the study for which the default R is calculated. (PDF) Click here for additional data file. 18 Jul 2021 Dear Dr Papadopoulos, Thank you very much for submitting your manuscript "Estimates of the basic reproduction number for rubella using seroprevalence data and indicator-based approaches" for consideration at PLOS Computational Biology. As with all papers reviewed by the journal, your manuscript was reviewed by members of the editorial board and by several independent reviewers. In light of the reviews (below this email), we would like to invite the resubmission of a significantly-revised version that takes into account the reviewers' comments. We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent to reviewers for further evaluation. When you are ready to resubmit, please upload the following: [1] A letter containing a detailed list of your responses to the review comments and a description of the changes you have made in the manuscript. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. [2] Two versions of the revised manuscript: one with either highlights or tracked changes denoting where the text has been changed; the other a clean version (uploaded as the manuscript file). Important additional instructions are given below your reviewer comments. Please prepare and submit your revised manuscript within 60 days. If you anticipate any delay, please let us know the expected resubmission date by replying to this email. Please note that revised manuscripts received after the 60-day due date may require evaluation and peer review similar to newly submitted manuscripts. Thank you again for your submission. We hope that our editorial process has been constructive so far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments. Sincerely, Benjamin Muir Althouse Associate Editor PLOS Computational Biology Nina Fefferman Deputy Editor PLOS Computational Biology *********************** Reviewer's Responses to Questions Comments to the Authors: Please note here if the review is uploaded as an attachment. Reviewer #1: The authors attempt to understand the variation in rubella’s basic reproduction number in countries for which pre-vaccination serological surveys are available in terms of measurable population characteristics. This is a thoughtful, carefully conducted, and important study with however disappointing results. Urban and rural differences would be expected by virtue of the relationship between contact rates and population density, and other patterns (e.g., education, healthcare, ...) probably are related. Other social phenomena should have greater impact than observed. Some thoughts: We know that the basic reproduction number is a property of a pathogen in a host population. One possible explanation for the as-yet-unexplained variation is that the pathogen is genetically heterogeneous (Rivailler et al. J Gen Virol 2017; 98:396-404). Another is that host age is far too influential to distinguish only children and adults, a possibility that the synthetic matrices of Prem et al. (PLoS Comput Biol 2017; 13:e1005697) make testable, assuming that mixing patterns haven’t changed. A third possibility is that the different assays (Dimech et al. Clin Microbiol Rev 2016; 29:163-74) that almost certainly were used obscured meaningful patterns. Reviewer #2: Thank you for the opportunity to review “Estimates of the basic reproduction number for rubella using seroprevalence data and indicator-based approaches.” The authors conduct a detailed analysis estimating the basic reproduction number for rubella in 98 different settings using three different methodological approaches. The objectives were to estimate R0 for rubella and to investigate the effectiveness of using prediction methods via linear regression and machine learning regression to predict R0 as compared to a simple regional averaging approach. The motivation for this paper builds on the intrinsic value of knowing the basic reproduction number for any infectious disease in order to understand transmission dynamics and importantly inform control measures. They find that rubella R0 is generally low (<5) and that neither random forest regression nor linear regression out performed a simpler regional averaging approach. Compared to other published papers, R0 is a little lower in this analysis, but still within the range typically estimated for rubella. One significant insight is that for countries missing seroprevalence data, a regional averaging approach may provide a reliable estimate of R0. The objectives are clearly laid out and addressed in the introduction and are original. The authors state that "the extent to which the regional average or particular indicators might predict what R0 might be for a given setting has not been studied for any infection." While other studies have evaluated associations between R0 and socio-demographic variables for other infectious disease, I am similarly unaware of any that evaluate the potential of particular indicators to predict R0 compared to other approaches. The methods are described in detail within the main text and supplemental information. The authors' use of multiple data sources that are publicly available is innovative and can be replicated for other (especially immunizing) infectious diseases. The methodology is rigorous and appropriate to answer the question at hand. Specially, the authors do a good job 1) including uncertainty in their estimates of R0 with bootstrapping and explaining within their limitations were uncertainly wasn't included and potential avenues for carrying over uncertainly in future work, and 2) exploring the impact of particular serological datasets on the correlation between indicators and R0. Lastly, the conclusions are consistent with the results presented. I have a four minor revisions and three potentially major revisions. Minor #1: Results first paragraph, description of Figure 1 findings. Please clarify where you are describing the country-specific R0 estimates, potentially specifying these as median estimates from the bootstrapped FoI estimates. It was just a little confusing to refer to the country-specific as point-estimates since the figure uses the term “point-estimate” for the regional estimates. Minor #2: Results second paragraph. I think the first sentence needs to be revised to say that R0 was higher (rather than lower) for urban than for rural regions. Minor #3: Results third paragraph, first sentence - Region of the Americas is listed twice. Minor #4: Page 10, second paragraph within section on correlation b/w R0 calculated using country-specific seroprevalence data and indicators - the sentence reads, “When the simple linear regression was repeated by region, economic indicators were the most correlated with the basic reproduction number in Africa (Table 2).” Is it rather... when the simple correlation analyses were repeated by region ...? Major #1: The broad findings from the work are based on the comparison of the prediction methods to the regional averaging (default) approach. I think the paper would benefit from a somewhat clearer understanding of the default approach. Specifically, what bootstrap-derived estimates of the force of infection are relied on in each analysis? In the non-imputed analysis described in S3 are the bootstrapped derived FoI estimates only from the settings in the three training folds, and does that explain why the MSE estimates vary across indicator in the non-imputed analysis compared to the imputed analysis? If true, are the FoI (from which to estimate a median R0) resampled 1000 times from the settings in the training folds, or are the FoI only a subset of the original 1000 bootstrapped samples? Also if true, how does this align with the discussion that states "The performance of this approach may have been overestimated given that bootstrap-derived estimates used to calculate the median would have included those from each of the studies being used to evaluate the performance," which reads as if bootstrap-derived estimates of the force of infection sampled from all 98 study sites were included in all default MSE estimates? Major #2: On a similar note, if indeed bootstrap-derived estimates of the force of infection sampled from all 98 study sites were included in all default estimates, this does seem to be a relatively systematic bias that would decrease the MSE. Is it possible to conduct a sensitivity analysis for at least one indicator or subset of indicators assessing R0 MSE if you didn't include FoI estimates from the study site you are trying to predict? Major #3: The authors state in the results that “For both assumptions about contact, estimates of R0 using the regional point estimate of the force of infection were usually smaller and led to a larger mean square error than the default R0 (Fig 1, Fig S2 and Table S5, Supplement),” but the significance of this finding is never revisited. It is possible to include a brief description of the motivation for this analysis and potentially the broad take away of this estimate under-performing compared to the regional averaging approach? ********** Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: No: The authors include references to data (that would need to be extrapolated), pseudocode, and in depth details of their methods, but do not actually include the data and code that can be used to easily replicate their findings. ********** PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No Figure Files: While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at . Data Requirements: Please note that, as a condition of publication, PLOS' data policy requires that you make available all data used to draw the conclusions outlined in your manuscript. Data must be deposited in an appropriate repository, included within the body of the manuscript, or uploaded as supporting information. This includes all numerical values that were used to generate graphs, histograms etc.. For an example in PLOS Biology see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5. Reproducibility: To enhance the reproducibility of your results, we recommend that you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option to publish peer-reviewed clinical study protocols. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols 12 Nov 2021 Submitted filename: ResponsesAndChanges.docx Click here for additional data file. 23 Jan 2022 Dear Dr Papadopoulos, We are pleased to inform you that your manuscript 'Estimates of the basic reproduction number for rubella using seroprevalence data and indicator-based approaches' has been provisionally accepted for publication in PLOS Computational Biology. Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow up email. A member of our team will be in touch with a set of requests. Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated. IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript. Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us now if you or your institution is planning to press release the article. All press must be co-ordinated with PLOS. Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Computational Biology. Best regards, Benjamin Althouse Associate Editor PLOS Computational Biology Nina Fefferman Deputy Editor PLOS Computational Biology *********************************************************** Reviewer's Responses to Questions Comments to the Authors: Please note here if the review is uploaded as an attachment. Reviewer #1: I have no additional comments. Reviewer #2: Thank you kindly for addressing my comments thoroughly. It is a very nice paper and I appreciate the opportunity to review it. ********** Have the authors made all data and (if applicable) computational code underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data and code underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data and code should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data or code —e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No 9 Feb 2022 PCOMPBIOL-D-21-00482R1 Estimates of the basic reproduction number for rubella using seroprevalence data and indicator-based approaches Dear Dr Papadopoulos, I am pleased to inform you that your manuscript has been formally accepted for publication in PLOS Computational Biology. Your manuscript is now with our production department and you will be notified of the publication date in due course. The corresponding author will soon be receiving a typeset proof for review, to ensure errors have not been introduced during production. Please review the PDF proof of your manuscript carefully, as this is the last chance to correct any errors. Please note that major changes, or those which affect the scientific understanding of the work, will likely cause delays to the publication date of your manuscript. Soon after your final files are uploaded, unless you have opted out, the early version of your manuscript will be published online. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers. Thank you again for supporting PLOS Computational Biology and open-access publishing. We are looking forward to publishing your work! With kind regards, Katalin Szabo PLOS Computational Biology | Carlyle House, Carlyle Road, Cambridge CB4 3DN | United Kingdom ploscompbiol@plos.org | Phone +44 (0) 1223-442824 | ploscompbiol.org | @PLOSCompBiol
  13 in total

1.  On the definition and the computation of the basic reproduction ratio R0 in models for infectious diseases in heterogeneous populations.

Authors:  O Diekmann; J A Heesterbeek; J A Metz
Journal:  J Math Biol       Date:  1990       Impact factor: 2.259

2.  The social contact hypothesis under the assumption of endemic equilibrium: Elucidating the transmission potential of VZV in Europe.

Authors:  E Santermans; N Goeyvaerts; A Melegaro; W J Edmunds; C Faes; M Aerts; P Beutels; N Hens
Journal:  Epidemics       Date:  2015-01-10       Impact factor: 4.396

3.  Modelling the incidence of congenital rubella syndrome in developing countries.

Authors:  F T Cutts; E Vynnycky
Journal:  Int J Epidemiol       Date:  1999-12       Impact factor: 7.196

Review 4.  The estimated mortality impact of vaccinations forecast to be administered during 2011-2020 in 73 countries supported by the GAVI Alliance.

Authors:  Lisa A Lee; Lauren Franzel; Jessica Atwell; S Deblina Datta; Ingrid K Friberg; Sue J Goldie; Susan E Reef; Nina Schwalbe; Emily Simons; Peter M Strebel; Steven Sweet; Chutima Suraratdecha; Yvonne Tam; Emilia Vynnycky; Neff Walker; Damian G Walker; Peter M Hansen
Journal:  Vaccine       Date:  2013-04-18       Impact factor: 3.641

Review 5.  Standardization of Assays That Detect Anti-Rubella Virus IgG Antibodies.

Authors:  Wayne Dimech; Liliane Grangeot-Keros; Christelle Vauloup-Fellous
Journal:  Clin Microbiol Rev       Date:  2016-01       Impact factor: 26.132

Review 6.  Using Seroprevalence and Immunisation Coverage Data to Estimate the Global Burden of Congenital Rubella Syndrome, 1996-2010: A Systematic Review.

Authors:  Emilia Vynnycky; Elisabeth J Adams; Felicity T Cutts; Susan E Reef; Ann Marie Navar; Emily Simons; Lay-Myint Yoshida; David W J Brown; Charlotte Jackson; Peter M Strebel; Alya J Dabbagh
Journal:  PLoS One       Date:  2016-03-10       Impact factor: 3.240

7.  Social contacts and mixing patterns relevant to the spread of infectious diseases.

Authors:  Joël Mossong; Niel Hens; Mark Jit; Philippe Beutels; Kari Auranen; Rafael Mikolajczyk; Marco Massari; Stefania Salmaso; Gianpaolo Scalia Tomba; Jacco Wallinga; Janneke Heijne; Malgorzata Sadkowska-Todys; Magdalena Rosinska; W John Edmunds
Journal:  PLoS Med       Date:  2008-03-25       Impact factor: 11.069

8.  Genetic diversity of currently circulating rubella viruses: a need to define more precise viral groups.

Authors:  P Rivailler; E Abernathy; J Icenogle
Journal:  J Gen Virol       Date:  2017-04-01       Impact factor: 3.891

9.  Projecting social contact matrices in 152 countries using contact surveys and demographic data.

Authors:  Kiesha Prem; Alex R Cook; Mark Jit
Journal:  PLoS Comput Biol       Date:  2017-09-12       Impact factor: 4.475

10.  The impact of Measles-Rubella vaccination on the morbidity and mortality from Congenital Rubella Syndrome in 92 countries.

Authors:  Emilia Vynnycky; Timoleon Papadopoulos; Konstantinos Angelis
Journal:  Hum Vaccin Immunother       Date:  2018-10-25       Impact factor: 3.452

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.