Literature DB >> 33951092

How well did experts and laypeople forecast the size of the COVID-19 pandemic?

Gabriel Recchia1, Alexandra L J Freeman1, David Spiegelhalter1.   

Abstract

Throughout the COVID-19 pandemic, social and traditional media have disseminated predictions from experts and nonexperts about its expected magnitude. How accurate were the predictions of 'experts'-individuals holding occupations or roles in subject-relevant fields, such as epidemiologists and statisticians-compared with those of the public? We conducted a survey in April 2020 of 140 UK experts and 2,086 UK laypersons; all were asked to make four quantitative predictions about the impact of COVID-19 by 31 Dec 2020. In addition to soliciting point estimates, we asked participants for lower and higher bounds of a range that they felt had a 75% chance of containing the true answer. Experts exhibited greater accuracy and calibration than laypersons, even when restricting the comparison to a subset of laypersons who scored in the top quartile on a numeracy test. Even so, experts substantially underestimated the ultimate extent of the pandemic, and the mean number of predictions for which the expert intervals contained the actual outcome was only 1.8 (out of 4), suggesting that experts should consider broadening the range of scenarios they consider plausible. Predictions of the public were even more inaccurate and poorly calibrated, suggesting that an important role remains for expert predictions as long as experts acknowledge their uncertainty.

Entities:  

Mesh:

Year:  2021        PMID: 33951092      PMCID: PMC8099086          DOI: 10.1371/journal.pone.0250935

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Expert opinion is undoubtedly important in informing and advising those making individual and policy-level decisions. In the early COVID-19 pandemic, clinicians, epidemiologists, statisticians, and other individuals seen as experts by the media and the general public, were frequently asked to give off-the-cuff answers to questions about how bad the pandemic might get. Answers to such questions draw upon “skilled intuition,” i.e., rapid judgments based on the recognition of similarities to other relevant situations, built up over a long period of experience [1, 2]—as well as the efficient recall of relevant information from long-term memory, which also benefits from expertise [3]. However, as the quality of expert intuition can vary drastically depending on the field of expertise and the type of judgment required [2], it is important to conduct domain-specific research to establish how good expert predictions really are, particularly in cases where they have the potential to shape public opinion or government policy. There has been limited research on the accuracy of expert and nonexpert COVID-19 forecasts, or the accuracy of levels of confidence in such forecasts (“calibration”). One study of COVID-19-related predictions from 41 U.S. experts found that their proposed method of aggregating predictions into a combined consensus distribution was more accurate than a more naïve way of aggregating expert forecasts which they referred to as an “unskilled forecaster” [4], but did not actually survey nonexperts or evaluate nonexpert predictions. Conversely, [5] investigated COVID-19 forecasts of nonexperts in Germany in the early pandemic, but did not investigate expert forecasts. One notable finding was that in a survey conducted in mid-March that asked participants to predict how many COVID-19 deaths would have occurred in Germany by the end of the year, the median estimate was exceeded just 16 days after the survey. We are not aware of any research specifically comparing expert and non-expert COVID-19 forecasts, although COVID-19 statistical and computational models and some of the expert forecasts based on them have been roundly critiqued for inaccuracy, overconfidence, and flawed assumptions [6-8]. It is important to differentiate between research evaluating the forecasts of ‘experts’—operationally defined in this paper as individuals holding occupations or roles in subject-relevant fields, such as epidemiologists and statisticians—and research evaluating specific epidemiological models, although expert forecasts may well be informed by epidemiological models. There has been more research evaluating the latter, which has found that many COVID-19 models achieve reasonable short-term predictions but that longer-term predictions are far more difficult due to the nonlinear nature of the processes that drive the spread of infection [9-11]. It is also important to distinguish between forecasts of arbitrary ‘experts’ as defined above, and those of the much smaller subset of experts who most directly inform public policy (e.g., scientific advisory committees, scientists participating in science policy initiatives, etc.); the latter are far harder to study, and we necessarily focus on the former here. Arguably, even the forecast accuracy of ‘experts’ by the broad definition is of some interest—should we put any more weight on an epidemiologist’s Twitter forecasts, or on those of someone interviewed by local media because they are a well-known statistician, than the predictions of a man or woman stopped on the street? In any case, systematic analysis of predictions for other viral outbreaks provides hints that we perhaps should not expect too much from either expert or nonexpert forecasts of COVID-19. For example, research conducted on disease forecasts (expert forecasts as well as model predictions) of the 2014 Ebola outbreak appearing in the published literature found that only 37% of predictions ended up being within 50% - 150% of the actual number of deaths [12]; furthermore, of seven predictions that were considered to represent “best case scenarios”, four of them still predicted death tolls higher than what ultimately resulted. As with the literature focusing specifically on COVID-19, evaluations of forecasts for other viral outbreaks have largely focused more on proposing model evaluation frameworks or evaluating specific models (e.g. [13, 14]) than on expert forecasts. However, to contextualize conclusions about expert predictions, it is critical to compare them to nonexpert predictions [15]. After all, if expert predictions are disregarded by the public, nonexpert predictions are liable to drive behavior in their stead. To this end, we conducted a survey of experts and nonexperts in April 2020, asking participants to make four predictions about the extent and severity of the COVID-19 outbreak by the end of 2020, and to indicate their confidence in their predictions by providing lower and upper bounds of an interval that they were 75% confident the true answer would fall within. (We refer to these as ‘75% confidence intervals’, following prior similar literature; this usage is clearly distinct from the traditional notion of a ‘confidence interval’ as an estimate computed from the statistics of observed data). The results provided clear evidence of differences between expert and nonexpert predictions in both accuracy and calibration.

Materials and methods

Participants were asked to make four COVID-19 forecasts (Table 1). In each case they were asked to estimate what the true answer to the question would be, and afterward to provide two additional numbers “in such a way that you think there’s about a 75% chance that the real-world answer will fall between your lower and higher number”. 75% was considered a reasonable value as it made it possible to identify how well-calibrated participants were on average: it left room for identifying both overconfidence (if fewer than three of the four real-world outcomes fell within an individual’s range) and underconfidence (if more than three did), and has been used in prior literature for similar reasons [16]. Nonexpert participants completed the survey as a part of a larger set of questions described elsewhere [17] which included the adaptive Berlin Numeracy test [18]. We surveyed 2,086 UK laypersons, sampling about half (N = 1047) from the survey platform Prolific Academic and the remainder from the ISO-certified panel provider Respondi.com, using proportional quota sampling to achieve a sample proportional to the UK population on age and gender. In parallel, we surveyed a convenience sample of experts recruited from social media. For the purposes of this survey, ‘experts’ were defined as epidemiologists, statisticians, mathematical modelers, virologists, and clinicians, as these represented the occupations/specialties of individuals commonly asked by the media to give predictions or expert opinions on COVID-19 in the early months of the pandemic. 140 respondents indicated that they resided in the UK and held one or more of these specialties.
Table 1

Questions asked of participants with corresponding forecast medians, median absolute deviation (MAD), median absolute error (MAE) and median relative error (MRE).

Question 1Question 2Question 3Question 4
QuestionHow many people in the country you’re living in do you think will have died from COVID-19 by December 31st 2020?How many people in the country you’re living in do you think will have been infected by COVID-19 by December 31st 2020?Out of every 1000 people who will have been infected by the virus worldwide, how many do you think will have died by December 31st 2020 as a result?Out of every 1000 people who will have been infected by the virus in the country you’re living in, how many do you think will have died by December 31st 2020 as a result?
How true outcome estimate was derivedTotal number of “deaths within 28 days of positive test” having a date of death earlier than 1 Jan 2021Number of infections implied by dividing the total number of COVID-19 deaths in the UK (left) by the UK infection fatality rate estimated by Imperial College COVID-19 response team in Oct 20201000 multiplied by the age-specific infection fatality rates estimated by the Imperial College COVID-19 response team in Oct 2020, weighted by worldwide age distribution1000 multiplied by the UK infection fatality rate estimated by the Imperial College COVID-19 response team in Oct 2020
True outcome estimate75,3466,385,2544.5511.8
Experts, median (MAD)30,000 (15,000)4,000,000 (3,687,500)10 (5)9.5 (4.5)
High-numeracy nonexperts, median (MAD)25,000 (10,000)800,000 (700,000)30 (20)30 (22)
All nonexperts, median (MAD)20,000 (10,000)250,000 (247,000)50 (45)40 (35)
Expert MAE45,3465,585,2545.456.80
High-numeracy nonexpert MAE55,3466,085,25425.4518.20
Nonexpert MAE55,3466,235,25445.4528.20
Expert MRE2.513.191.982.03
High-numeracy nonexpert MRE3.327.985.593.20
Nonexpert MRE3.7725.549.193.98
All participants completed the survey between 7 April and 12 April 2020, with the exception of 5 experts who submitted responses between 14 and 16 April, one who submitted a response on 20 April, and one who submitted a response 7 May. The expert sample contained 19 epidemiologists, 65 statisticians, 44 mathematical modelers, 35 clinicians, and 1 virologist; this was defined as the list of “expert occupations” prior to survey distribution, and data from experts who did not hold one of these occupations/roles was excluded. (Numbers in the previous sentence do not sum to 140, as some individuals reported holding multiple roles). Answers to survey questions that were entered in non-machine-readable ways (e.g., the phrase “1 million” rather than “1,000,000”) were normalized with a combination of regular expressions and manual inspection. Removing blanks, uninterpretable answers, answers exceeding 1000 (Q3 & Q4) or the population of the UK (Q1 & Q2), and answers which did not follow instructions yielded 405 point estimates and 402 ranges from experts, and 7,593 point estimates and 6,801 ranges from nonexperts. Free text responses highlighted that some nonexpert participants found the instructions difficult to understand or follow, so we restricted our primary nonexpert analysis to those who scored in the top quartile of numeracy (N = 524), and reserved the full nonexpert pool for a secondary analysis. Accuracy and calibration were calculated by comparing participants’ April estimates to a “true outcome estimate” determined in January 2021. Specifically, the total number of COVID-19 deaths by December 31 (Question 1) was assessed using the official criterion of the United Kingdom: the total number of “deaths within 28 days of positive test” having a date of death earlier than 1 Jan 2021, as reported at https://coronavirus.data.gov.uk/details/deaths [19]. The total number of infections by December 31 (Question 2) was estimated by computing the total number of COVID-19 deaths in the UK as of December 31 by the UK infection fatality rate (IFR) estimated by the Imperial College COVID-19 response team in October 2020 [20]. The true outcome estimate for Question 4 was obtained by multiplying this same estimated IFR by 1000. Finally, the true outcome estimate for Question 3 was obtained by multiplying the age-specific IFRs estimated by the Imperial College COVID-19 response team [20]—see [20]’s Table 2—weighted by the worldwide age distribution [21]. In other words, the IFR for 25–29 year-olds was multiplied by the proportion of the world population aged 25–29, the IFR for 30–34 year-olds was multiplied by the proportion of the world population aged 30–34, and so on, with the sum of the results for all age brackets treated as the world IFR. As best we can determine from the description available, this was the method used by the Imperial College team to estimate fatality rates across countries of different income brackets, but applied to the world rather than only those countries in a specific income bracket. These outcome estimates necessarily remain approximate but were presumed adequate to compare expert and nonexpert predictions.
Table 2

Proportions of participants from each group (experts, high-numeracy nonexperts, and all nonexperts) for whom the outcome fell within their own 75% confidence intervals.

Question no.ExpertsHigh-numeracy nonexpertsAll nonexpertsX2, experts vs. high-numeracy nonexpertsX2, experts vs. all nonexperts
(p-value)(p-value)
139/108 (36%)78/483 (16%)169/1757 (10%)22.2 (p < .001)72.1 (p < .001)
240/100 (40%)58/479 (12%)133/1737 (8%)45.8 (p < .001)115.9 (p < .001)
341/98 (42%)47/466 (10%)159/1634 (10%)62.0 (p < .001)93.3 (p < .001)
455/96 (57%)129/474 (27%)330/1673 (20%)33.0 (p < .001)75.2 (p < .001)
This study was overseen by the Psychology Research Ethics Committee of the University of Cambridge (approval number PRE.2020.034, amendment 1 April 2020). Participants viewed a Participant Information Sheet and provided written consent via an electronic consent form before accessing the survey. Data and code are available at https://osf.io/dcn5q. See S1 Appendix for questionnaire items.

Construction of linear opinion pools (consensus distributions)

In addition to calculating the accuracy and calibration of experts and nonexperts as individuals, consensus distributions were generated for each question by aggregating distributions having 75% of the probability mass uniformly distributed within each participant’s 75% confidence interval. This required constructing and combining distributions for individual participants, and enabled us to calculate continuous ranked probability scores (CRPS), a common approach to comparing probabilistic forecasts. Although we computed the CRPS of aggregated consensus distributions, if we were to compute it for an individual forecast (after transforming that forecast to a probability distribution in the manner described later in this section), it would most reward participants with relatively narrow 75% confidence intervals that also contained the correct outcome; it would give poorer scores to participants whose 75% confidence intervals contained the correct outcome, but were somewhat wider, and would give especially poor scores to participants with narrow 75% ranges that were not anywhere close to the correct outcome. CRPS is therefore useful in this study as it serves as a measure of accuracy that can be applied to a consensus distribution constructed from participants’ 75% confidence intervals, rather than from their point estimates. An alternative, perhaps more straightforward approach would have been to use the weighted interval score [22], which approximates the CRPS and does not require the construction of a full predictive distribution. Constructing consensus distributions also allowed us to better visualize the aggregated predictions. For each of the four questions, a separate consensus distribution was constructed for experts, nonexperts, and high-numeracy nonexperts, yielding a total of twelve consensus distributions (Fig 1). These distributions assigned zero probability mass to values less than 0 or values greater than a “natural maximum”: 1000 for questions 3 & 4, which asked participants to estimate how many people out of every 1000 would experience a particular outcome; and 67 million (the approximate population of the UK) for questions 1 and 2, which asked participants to estimate how many people in the UK would experience a particular outcome. Probabilities were assigned to outcomes between 0 and the natural maximum by averaging together probability distributions constructed for each participant. The distribution for each [expert / high-numeracy nonexpert] was constructed as follows: 75% of the probability mass was distributed uniformly within the interval [lower, upper], where lower and upper refer to the lower and upper bounds of the range provided by the participant; the width of this interval is given by upper–lower. The remaining 25% was distributed uniformly within [min, lower) ∪ (upper, max], where min was the larger of 0 and lower–(width/2), and max was the smaller of upper + (width/2) and the natural maximum. This had the effect of restricting the total width of each individual distribution to a width equal to or less than twice the width of the range provided by the participant, and was done in order to ensure that the remaining 25% was allocated to values close to the participant’s 75% confidence interval. While the specific choice of how much to restrict the distribution was necessarily somewhat arbitrary, if we were to assume that the participant’s 75% confidence interval corresponds to the “middle 75%” (i.e., the 12.5th to 87.5th percentiles) of a Gaussian distribution, two times this width would span from the 1st percentile to the 99th percentile. It therefore seems likely that an individual who is truly 75% confident that a value will fall between lower and upper would be extremely surprised if the value was less than lower–(width/2) or higher than upper + (width/2). Restricting the distributions within these bounds therefore seemed justified. Finally, probability distributions of different individuals were combined with simple averaging of the probabilities (i.e. ‘vertical’ combination).
Fig 1

Consensus distributions (linear opinion pools) for Q1 (a), Q2 (b), Q3 (c), and Q4 (d). Axes truncated to allow the overall shapes of the distributions to be visible.

Consensus distributions (linear opinion pools) for Q1 (a), Q2 (b), Q3 (c), and Q4 (d). Axes truncated to allow the overall shapes of the distributions to be visible. For nonparametric distributions, the scoringRules R package [23] can make use of random samples from the forecast distribution to approximate the CRPS. We randomly sampled sets of 5,000 samples from the consensus distributions repeatedly (100 times per distribution), resulting in 100 approximate CRPS scores for each consensus distribution.

Results

Accuracy

As reported in Table 1, in terms of absolute error, expert point estimates for each of the four questions were more accurate than those of high-numeracy nonexperts. Mood’s median tests indicated that these differences were significant: Q1, p = .03, Q2, p = .04, Q3, p < .001, Q4, p < .001. Expert point estimates were also more accurate than those of all nonexperts, Q1, p < .001, Q2, p = .003, Q3, p < .001, Q4, p < .001. Similar results emerged when evaluating relative error rather than absolute error. Several measures of relative error exist; for ease of interpretation we here report the exponential of the absolute value of the log difference measure [24], ln( / x). This is a natural measure for predictions of this kind as it is scaled to the size of the true outcome. For example, a prediction that is 1/3 the size of the true outcome is treated as having the same amount of relative error as a prediction that is 3 times the size. As shown in Table 1, experts had less relative error than nonexperts, with their medians being lowest for each question. Mood’s median tests indicated that differences in relative error between experts and high-numeracy nonexperts were not significant for Q1 (p = .07) but were significant at p < .001 for the remaining three questions. Differences in relative error between experts and all nonexperts were significant at p < .001 for all questions.

Calibration

Calibration could be assessed for the 93 experts, 1459 nonexperts, and 434 high-numeracy nonexperts who fully answered all questions with clearly interpretable responses. Given that four questions were asked, the number of outcomes falling within an optimally calibrated individual’s 75% confidence interval has an expected value of three. Experts came closer to meeting this standard than nonexperts: The median number of outcomes falling within each participant’s interval was 2 for experts (mean = 1.82, SD = 1.17), 0 for high-numeracy nonexperts (mean = 0.68, SD = 0.90) and 0 for all nonexperts (mean = 0.49, SD = 0.77). Mood’s median tests indicated that the differences in medians between experts and nonexperts, and between experts and high-numeracy nonexperts, were significant, both p < .001. 20 of 93 experts (22%), 33 of 1459 nonexperts (2%), and 16 of 434 high-numeracy nonexperts (4%) were calibrated such that exactly 75% (three) of the four outcomes fell within their 75% confidence intervals. We also calculated the proportion of participants from the given group (experts, high-numeracy nonexperts, or all nonexperts) for whom the outcome (“true outcome estimate”) fell within the participant’s 75% confidence interval, inclusive; other investigators have treated this proportion as a measure of ‘calibration of confidence assessments’ for a group [25]. For a group in which each individual was perfectly calibrated, 75% of participants’ 75% confidence intervals would be expected to contain the true value. The “true outcome estimate” fell within the 75% confidence intervals for Q1 to Q4 of 36%, 40%, 42%, and 57% of experts respectively, but within the ranges of only 16%, 12%, 10%, and 27% of high-numeracy nonexperts and 10%, 8%, 10%, and 20% of all nonexperts. Chi-squared tests indicated that these proportions differed significantly between experts and high-numeracy nonexperts on every question, and that differences between experts and all nonexperts were likewise significant (Table 2).

Continuous ranked probability scores

Experts had lower (i.e., better) CRPS than high-numeracy nonexperts on each question, indicating superior forecasting overall. Although it was clear that this was the case from the descriptives alone, two-tailed t-tests were nevertheless used to formally test whether the mean of the 100 approximate CRPS scores calculated from the expert consensus distributions was indeed different than that of those calculated from the high-numeracy nonexpert consensus distributions (Table 3). As the sampling procedure to obtain the inputs for the CRPS analysis was computationally intensive and time-consuming, and as the full nonexpert distributions clearly predicted outcomes as poorly or more poorly than the high-numeracy nonexpert distributions (Fig 1), it was not deemed necessary to repeat this analysis for the full nonexpert consensus distributions.
Table 3

Descriptive and inferential statistics for the sets of 100 approximate continuous ranked probability scores generated from expert and high-numeracy nonexpert consensus distributions.

Question no.ExpertsHigh-numeracy nonexpertsTwo-tailed t-test comparing experts and high-numeracy nonexperts
MeanSDMeanSDtdfp
124,30128931,301364150.66188.27< .001
23,210,15363,5853,563,70232,72149.44148.00< .001
37.440.1726.600.60306.35115.06< .001
43.460.0720.270.54310.51102.50< .001

Accounting for demographic differences

Experts had a mean age of 42.3 (95% CI 40.0–44.7), slightly younger than nonexperts’ mean age of 45.3 (95% CI 44.6–46.1). Experts were also 75% (66% - 83%) male; this proportion was higher than either the high-numeracy nonexpert group—51% (47% - 56%) male—or the group of all nonexperts, 48% (46% - 51%) male. When regressing rank-transformed absolute error on gender, age, and expert/nonexpert status, expert status was significantly associated with lower error for each of the four prediction questions, with betas for expert status being 2 to 16 times the magnitude of the corresponding beta for gender, depending on the question (Table 4). Older age was associated with (lower) error for question 4 only. Similarly, expert status (β = 1.3) and male gender (β = 0.2) were both significantly associated with the number of outcomes falling within each participant’s 75% confidence interval.
Table 4

Regressions of calibration and accuracy on gender, age, and expert/nonexpert status.

Predictorβ
# outcomes within rangeQ1 errorQ2 errorQ3 errorQ4 error
Expert status1.284***-352.22***-177.61**-633.91***-588.34***
Age-0.0010.061-0.592-0.545-1.658*
Male gender0.162***-22.22-73.68**-102.21***-82.76**
Adjusted r20.1450.0130.0060.0530.045

Note. ‘Error’ represents rank-transformed absolute error. Stars represent significance at p < .05 (*), p < .01 (**), p < .001 (***).

Note. ‘Error’ represents rank-transformed absolute error. Stars represent significance at p < .05 (*), p < .01 (**), p < .001 (***). Numeracy could not be included in these regressions as we did not require experts to complete the Berlin numeracy test. However, when these regressions were repeated but with nonexperts restricted to the subset of high-numeracy nonexperts, expert status continued to significantly predict the number of outcomes falling within each participant’s interval, and also continued to significantly predict rank-transformed error for each of the four questions, with the exception of question 2. Restricting the dataset to only nonexperts allowed numeracy to be included as a predictor along with age and gender. Higher numeracy was significantly associated with more outcomes falling within each participant’s 75% confidence range, as well as with lower error on each prediction question. Male gender remained associated with more in-range outcomes, and also remained associated with lower error on three of four prediction questions, but the beta for gender had lower magnitude than the beta for numeracy in each case. As before, age was associated with error for question 4 only.

Discussion

Despite the limitations of this survey, there are nevertheless a few key lessons to be drawn. First, the experts in our study demonstrated overconfidence: out of the four intervals that experts expected outcomes to fall within 75% of the time, fewer than half of actual outcomes fell within these intervals on average. This is perhaps unsurprising given reports of poor calibration of disease models of COVID-19 and the 2014 Ebola outbreak [8, 12], but it is noteworthy that this was true even when experts were being asked to fill out an informal survey—a context in which most experts presumably did not run their favorite epidemiological model to see what it predicted by year end. In the present case, numbers of deaths and infections by the end of the year were substantially more severe than most expert predictions, unlike in the 2014 Ebola outbreak, when outcomes were less severe than predicted by experts [12]. The common theme seems to be that estimates of the likely intervals in which future observations would fall were too narrow. Second, nonexpert predictions were less accurate than expert predictions, and nonexperts were more overconfident than experts in their predictions. This was true even of those nonexperts who scored in the top quartile of a standard test of numeracy. Follow-up analyses suggested that these differences were not due to confounds with age or gender. Therefore, although our findings on expert accuracy and overconfidence may read as a cautionary tale against taking expert predictions at face value, it is critical to highlight that we could do worse: we could believe the predictions of people who are not experts. We have arguably witnessed many examples of the latter approach being taken by individuals across the globe, sometimes with dire results. As [15] notes, the (in)accuracy of lay predictions is essential context when discussing expert performance. Focusing solely on poor expert performance may simply make nonexperts more adamant about their own preconceptions—not a good thing if they are already even more inaccurate and more overconfident than the experts, as our results suggest. A key limitation with respect to what this study can tell us about expert predictions in the real world is that there is enormous heterogeneity both among experts and the conditions under which they make forecasts. The concept of an ‘expert’ as operationalized in this study is extremely broad. It would be especially helpful to understand more about the accuracy of forecasts produced by the subset of experts who are most influential, e.g., those who sit on scientific advisory committees that inform policy. These are a very specific subset of ‘experts’ who are harder to study, and it is an open question to what degree any findings about off-the-cuff predictions of the ‘experts’ recruited in our study might generalize to predictions made by those experts who have an explicit mandate to produce forecasts for policymakers. Furthermore, there are ‘experts’ who are selected as policy advisors (or as interviewees on media programs) because it is known that they support policies preferred by a particular political party; experts who choose to become involved in policy precisely because they feel more confident than others about what policies should be advocated for; experts who may be incentivized to be more cautious, fearing that reputation and employment opportunities may suffer if they get things too badly wrong; and experts who thrive on media attention (or are even employed as media pundits), who may have more incentive to make attention-grabbing predictions than accurate ones. In other words, individual incentives and incentive structures likely have enormous influence on accuracy and calibration, and it would be a mistake to assume that our results are equally applicable to all subtypes of ‘expert’. However, there are some reasons to believe that the general finding of overconfidence (among both ‘experts’ and laypeople) is likely to generalize to a number of contexts in which COVID-19 forecasts (and presumably epidemiological forecasts more generally) are made. First, the finding of overconfidence among individuals with relevant subject-matter expertise is consistent with, in the words of Philip Tetlock, “a multi-decade line of psychological research on expert judgment that has shown that experts in a wide range of fields are prone to think they know more than they do (the overconfidence effect)” [26]; other researchers have described overconfidence as “the most ubiquitous bias in studies of calibrated judgments about risks and uncertainties” [27]. Research summarizing relevant studies across a wide variety of fields finds systematic overconfidence in judgments made by both lay predictors and those with relevant expertise [28-30]. This is not solely a laboratory phenomenon. For example, in an analysis of roughly five thousand 90% confidence ranges from 27 studies mostly “performed in the course of applied research in the experts’ domains of expertise, not primarily in laboratory studies of overconfidence,” and where all respondents were “experienced professionals giving judgments on important real-world problems within their own domains of expertise,” Lin and Bier [31] found much variation in the extent of overconfidence, but found overconfidence overall. On average, the percentage of true values falling within respondents’ 90% probability intervals was less than 90% for each of the 27 studies they analyzed, although when digging deeper into the questions asked by each study, they found that the level of overconfidence was highly variable and that “some questions even produced under-confidence.” This is not to say that overconfidence in judgments about future events is universal among experts or nonexperts. Overconfidence is influenced by the frequency and kind of feedback that individuals typically receive about their predictions; for example, weather forecasters appear to be particularly well-calibrated [28]. It also seems highly likely that forecasting experience and the incentive structure of the social environment in which the forecasts are made will have an effect. For example, members of the Good Judgment Project, a group of forecasters known for their exceptional performance in a multi-year geopolitical forecasting competition conducted by the U.S. Intelligence Advanced Research Projects Activity, achieved exceptionally high calibration [32]. They attributed this success to the fact that they explicitly aimed to "structure the situation, incentives, and composition" of their team so as to produce accurate and well-calibrated forecasts, and also present evidence that training and interaction in teams was beneficial. We therefore certainly would not conclude that domain experts untrained in forecasting would perform better than large groups of ‘nonexperts’ who are practiced in forecasting future events. One such group appears to have outperformed experts in infectious disease modelling on multiple COVID-19-related forecasts, on the forecasting platform Metaculus [33]. Promising follow-up research is beginning to combine the predictions of ‘nonexpert’ forecasters from Metaculus and the Good Judgment Project with those of epidemiological modelers to produce consensus forecasts of hopefully greater accuracy than either in isolation, as well as a ‘meta-forecast’ which combines this consensus forecast with an ensemble of forecasts from computational models [34]; the results have yet to be systematically evaluated. Other initiatives to solicit and evaluate a wide range of approaches to epidemiological forecasting, such as the DARPA Chikungunya challenge [35], in combination with research on approaches to aggregating forecasts of subject-matter experts [36] and nonexperts [37-39], have also established promising routes toward improving forecasting of epidemics. In other words, we are not all doomed to be overconfident: there is much that can be done to improve the accuracy and calibration of forecasts, at least in the context of forecasting tournaments. Forecasting experience and the incentive structure of the social environment are likely to affect predictions made outside of forecasting tournaments as well. For example, Mandel and Barnes’ analysis [40] of over 3,500 geopolitical forecasts from intelligence analysts, extracted from real-world reports, found systematic underconfidence rather than the overconfidence seen in Tetlock’s famous studies of expert geopolitical forecasting, or the excellent calibration seen in forecasting tournaments. Although there were many differences that may have contributed to the different findings, the researchers noted one plausible factor was that “accountability pressures on analysts are far greater than those placed on forecasters in geopolitical tournaments,” citing research from Tetlock & Kim showing improved calibration in the presence of social accountability [41]. Incentives to be accurate and well-calibrated in our study were low for experts and nonexperts alike. It is possible that, with respect to public predictions which may influence individual behavior, some experts may have an especially strong incentive to be as correct and cautious as possible, feeling that their reputations are on the line and that their predictions will be subject to future scrutiny. Intuitively, this seems likeliest to be true for experts who feel that they benefit much more from a reputation for accuracy than from media attention—although these experts may also be least likely to voice public predictions about highly uncertain events in the first place. Unfortunately, however, for many public predictions, there are reasons to believe the incentive structure may be the other way round. For example, Tetlock and colleagues asked expert participants “how often they advised policy makers, consulted with government or business, and were solicited by the media for interviews”, reporting a significant positive correlation between these assessments and the degree of overconfidence exhibited by these experts in geopolitical predictions [42]. They also reported a positive correlation between overconfidence and Google search counts (used as a proxy for the number of times participants were mentioned in the media). Tetlock points out that the causal links may be bidirectional: “On one hand, overconfident experts may be more quotable and attract more media attention. On the other, overconfident experts may also be more likely to seek out the attention” ([42], p. 63). Either way, the result is the same: predictions made by overconfident experts may be the most visible. In sum, our findings of overconfidence in lay and expert COVID-19 predictions are consistent with what would be expected from literature on predictions and judgments in other domains. They seem likely to generalize to common real-world contexts in which everyday people encounter expert or nonexpert predictions (e.g., within traditional or social media), but not necessarily to all such contexts. They also may not generalize to contexts where accuracy and good calibration are strongly incentivized, or in which individuals receive systematic training or practice in forecasting with regular feedback. Are there other data suggesting overconfidence in experts or nonexperts of COVID-19 case count and death prediction? There exist a very interesting set of reports associated with the study mentioned in the introduction by McAndrew and Reich, which solicited COVID-19-related predictions from U.S. experts in a total of twelve surveys. The fifth survey (administered March 16–17, 2020), the first to ask both about the ‘smallest, most likely, and largest’ number of U.S. COVID-19 deaths they expected by the end of 2020, as well as about the ‘smallest, most likely, and largest’ number of U.S. COVID-19 cases that they expected to have been reported for specific dates in the future, was completed by 18 experts in infectious disease modelling [43]. Although the data for individual expert predictions is not available, the 80% confidence interval of the experts’ consensus distribution for reported U.S. COVID-19 cases by March 22nd was reported as 7,061–24,180 [43], and the 80% confidence interval for March 29th was 10,500–81,500, according to a news article appearing on ABC News’ FiveThirtyEight about the same survey [44]. The ‘smallest, most likely, and largest’ estimates of each expert for the latter prediction were also visualized in the article. Ultimately, according to the ‘truth database’ later compiled by the survey authors [45], the true outcomes for the corresponding questions were 33,404 and 139,061, respectively, both well outside the 80% confidence intervals. From FiveThirtyEight’s visualization, it can be inferred that the true March 29th count exceeded the high-end estimate of 15 of the 18 modellers. On the other hand, the expert consensus 80% confidence interval for the number of U.S. COVID-19 deaths in 2020 was 195,000–1.2 million, with a point estimate of 195,000; the CDC’s ultimate 2020 COVID-19 death count of 379,705 [46] fell within this range, although it exceeded the high-end estimate of 240,000 announced by U.S. administration scientists on 31 March 2020 [47]. By the time of McAndrew and Reich’s final survey, administered May 4–5, they were also soliciting crowdsourced predictions on the ultimate number of U.S. COVID-19 deaths expected by the end of 2020 from Metaculus; by this time, experts were predicting a median 256,000 deaths (80% CI 118,000–1.2 million), and Metaculus predictors a median 197,000 (80% CI 120,000–460,000) [48]. It is difficult to draw too many conclusions from these examples, but they do provide an example of overconfident COVID-19 predictions even among infectious disease modelers, making predictions for just one week into the future, early in the pandemic. The non-linear dynamic nature of infectious diseases makes possible futures especially uncertain–small initial differences in infection parameters can lead to big differences in outcomes with time–and it certainly seems plausible that this makes it challenging to estimate one’s own level of certainty. However, overconfidence is not necessarily restricted to experts or to epidemiological forecasts, as the prior literature we have discussed makes clear. We also expect that our finding of poorer accuracy in COVID-19-related predictions for the “person on the street” versus people with relevant subject-matter expertise is likely to generalize to some degree, but not to all nonexperts. In addition to being a common-sensical finding, ‘relevant subject-matter expertise’ typically includes knowledge of what real-world data sources contain the most reliable information; for example, it seems plausible that our ‘experts’ did well on questions 3 and 4 simply because they were aware of preliminary estimates of the infection fatality rate, whereas our nonexperts may not have been. Individuals with professional backgrounds in statistics, mathematical modelling, and epidemiology also seem likelier to have had a greater familiarity with the raw data around COVID-19 infections and death rates, a firmer grounding in how to interpret that data, and the high levels of uncertainty associated with epidemiological forecasts in general. As mentioned, however, the impressive performance of practiced forecasters on crowdsourcing platforms suggests that domain expertise is not necessarily a prerequisite for good forecasting. Other limitations of this study include convenience sampling of experts and a small number of correlated questions. Participants were also asked to produce point estimates before they were asked to produce ranges, which is known to anchor responses toward the point estimate [49], so participants may have seemed more overconfident than they would have with other elicitation methods. Nevertheless, given the stark differences between expert and nonexpert accuracy and calibration levels, it seems unlikely that alternate elicitation methods would erase these differences. Additionally, we purposefully cast a broad net in terms of our recruitment of experts, and it is certainly possible that a more rigorous process for selecting individuals with the most relevant subject-matter expertise would have resulted in a set of experts who made more accurate predictions. However, a more careful selection of experts would presumably have simply made the differences between nonexpert and expert performance even more stark (if it had any effect at all), rather than the reverse, so there seems little reason to question the performance differences between experts and nonexperts.

Conclusions

Much of the discussion around communicating forecasts in the COVID-19 pandemic has centered around tradeoffs in communicating uncertainty with respect to public trust. For example, in some contexts downplaying uncertainties may shore up public trust in the short term, but confident predictions that later turn out to be wrong may reduce public trust in science [50]. Overall, our results underscore a need for individuals with expertise in fields relevant to forecasting epidemiological outcomes (and who communicate about these forecasts publicly) to consider broad ranges of possible outcomes as plausible, and to consider communicating this high level of uncertainty to nonexperts. The ultimate message may be that “the experts have much to learn, but they also have much to teach” [15]. Given the continued impact of COVID-19 and risks of other future pandemics, further research into improved epidemiological forecasting may prove vital. In the meantime, we must all learn to acknowledge and admit that the uncertainties may be greater than we think they are, whether we are experts or not.

Questionnaire items.

Expert and nonexpert questionnaire items. (DOCX) Click here for additional data file. 20 Mar 2021 PONE-D-21-04714 How well did experts and laypeople forecast the size of the COVID-19 pandemic? PLOS ONE Dear Dr. Recchia, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by May 04 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript: A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols We look forward to receiving your revised manuscript. Kind regards, Martial L Ndeffo Mbah, Ph.D Academic Editor PLOS ONE Additional Editor Comments: This is a very interesting manuscript addressing a timely question of disease forecasting. Though reviewers mainly pointed out some minor issues that need to be addressed, providing more clarification on the use of terms 'expert' and 'non-expert' will greatly improve the quality and potential impact of the study. To this effect, I fully agree with Reviewer #1 suggestions, and strongly encourage you to address them thoroughly. Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: No ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Summary While expert opinions have exerted a large influence on public policy during the Covid-19 pandemic, relatively little research has been done that examines the accuracy of such expert predictions. To fill that gap, the paper shows the results of a survey done in April on experts and non-experts who were asked to submit a prediction for four different covid-related targets. The paper makes a compelling case that experts (and to an even greater extent non-experts) have a tendency to be overly certain in their estimates and underestimate the possibility of extreme events. It is interesting to see that this tendency was different between experts and non-experts. It is also interesting that off-the-cuff predictions, on average, probably are not helpful to inform public policy. Major issues The idea of comparing experts and non-experts is in principle very interesting - especially given the importance of expert opinion for shaping public policy. One factor that makes studying this subject difficult, however, is the immense variability hidden behind the terms 'expert' and 'non-expert' that is (naturally) hard to capture. For example, a scientific policy advisor may be a quite different type of expert (with potentially much better forecasting skills) than someone who simply has a statistics background. Similarly, a layperson putting a lot amount of effort into forecasting questions on Metaculus, or someone who is a Superforecaster (see https://goodjudgment.com/) may make forecasts very differently than the average person who would participate in a study through Prolific Academic or Respondi. One way in which I especially fear study participants may differ from either 'experts' who inform policy or voice predictions publicly, or 'non-experts' who make public predictions is 'skin in the game' and the incentive structure they are faced with. Experts who make public predictions and who take responsibility may have very strong incentives to be as correct as possible (and possibly as cautious). Even if they make off-the-cuff prediction, they will probably be aware of public scrutiny. Similarly, laypeople who predict to earn money or to rank highly on a public leaderboard also have an incentive to be right. This likely greatly enhances predictive accuracy (see e.g. literature on Superforecasters). Experts surveyed in the study however, as I understood, had no real incentive to be right (except for the fact that they were recruited as 'experts' and therefore possibly felt some psychological pressure to live up to that standard). Non-experts presumably had no incentive at all to be right and maybe even had a financial incentive (in terms of their hourly rate) to complete the online survey quickly. The authors briefly mention how some people on Metaculus may be outperforming experts consistently, but unfortunately do not address this further or contextualise it in light of their research. I feel it would be helpful if the authors were more explicit on these limitations or address them otherwise. They do state unequivocally how they recruited their sample of 'experts' and 'non-experts', but could be clearer on the implications of that selection process and the study design. It would help to expand on how comparable the 'experts' in the study are to the 'experts' that actually do inform public policy. In the paper the same term 'expert' is used in two ways: First in the introduction and conclusion that talks about how important expert opinion is for public policy, and secondly to denote the group of people in the study with a background in certain fields. To the reader, this implies some comparability. I think it needs more explanation to argue that forecasts from the former group are actually comparable to predictions from the latter - or an explicit statement that they may not be. One option could be to call them 'study participants with a background in subject relevant fields' (or probably a catchier version of that) in the paper. Similarly, I think it should be addressed that the group of laypeople in the study may not be able to represent all kinds of 'non-experts' predictions, or even the ones we might care about most (e.g. Metaculus, Predictit etc.). That being said I do find it intuitively plausible that all experts, including those making public predictions have a tendency to be overconfident, even if they are incentivised to be right. I just believe this link needs more supporting evidence. One way to do show this could be to check for explicit predictions made by experts around the same time and see whether they are comparable with what the study participants predicted. Minor issues . line 66 page 3 the reporting of the Ebola study is worded slightly confusingly. What is 4 of 7? - line 123 I believe it should be 'divided by 1,000' instead of multiplied by 1,000 - In general, I feel it is somewhat unfortunate that for 3 out of 4 questions, there is no actual "truth data", but instead only estimates. This should ideally be thought of beforehand for future studies - Specifically for question 3 (lines 123ff) it is unclear to me what was multiplied with what to obtain death numbers. I'm also not quite clear why the authors did not simply take publicly reported death numbers? - Regarding the construction of linear opinion pools: the authors could maybe be clearer on a) why they chose to construct distributions for every individual expert at all. Was this just for the purpose of scoring, or was this mainly to construct an overall aggregate distribution? b) how their choice of restricting the distribution with 'width/2' was motivated? c) how exactly they aggregated individual expert distributions to one distribution. Did they combine the CDFs horizontally, i.e. use a quantile average, or vertically, i.e. construct a mixture distribution? - Regarding the scoring of forecasts it could be made clearer why the authors used the crps instead of weighted interval score (e.g. Bracher et al. 2021 https://doi.org/10.1371/journal.pcbi.1008618). It seems much more intuitive to calculate the wis for every individual expert and average it for every subgroup (experts, non-experts etc) the authors are interested. This is much easier than constructing a full predictive distribution and using the crps to score it. A software implementation of the WIS is for example available in the scoringutils package (https://github.com/epiforecasts/scoringutils - Disclaimer: I'm an author of that package). - Regarding the assessment in terms of relative scores in lines 176ff: I am unsure whether the relative assessment is actually necessary, as the main message is already clear in terms of absolute differences and so a relative assessment doesn't seem to add much. - stylewise, the reporting of the numerical results in the sections "Calibration" and "Continuos ranked probabilty" is somewhat hard to follow as there are a lot of numbers, maybe a table would be better suited to present the different p-values and tests performed? Especially in the section on the crps, the authors could be clearer on what numbers they report (for example, I'm not sure what the negative numbers mean). - A minor point regarding the assessment of calibation in lines 188ff: From the text I understand that the authors assume that a forecaster whose prediction intervals cover more of the true values are better calibrated. While this seems true in the case where all participants are overly confident in their predictions, it doesn't hold in general: a forecaster who covers 100% of the observed values with his prediction intervals (instead of the 75% aimed for) is not necessarily better calibrated than one who is right 50% of the time. Reviewer #2: The paper analyzes the precision of forecast regarding the magnitude of the COVID-19 pandemic by experts and non-experts. They compare the forecast in spring 2020 at the initial stage of the pandemic with the outcomes in December in the same year. They found that experts perform better in both accuracy of prediction and accuracy of levels of confidence than non-experts. The research addresses an important policy-relevant question. Major comments. 1. They compare raw data across groups. However, individual characteristics, such as gender or age, may explain some of the observed difference. Please compare the distribution (or mean) of individual characteristics across the three groups. I suspect the distribution is not necessarily similar given their sampling scheme. If it is the case, is there any way to control for the difference in some of the analyses? Comments. 2. P4. L86.There are two ways to survey non-experts: Prolific Academic and Respondi.com. What’s the reason for using two ways for survey? Is there any difference in respondents’ characteristics and/or responses? Please explain. 3. P4. L84. How do they pick up the particular value of 75% for calibration? Is this the common number used in literature? 4. P9. L162. Please explain what CRPS shows in the method. It seems it analyzes the accuracy of forecasting, but I would like to know more details on why these analyses are necessary. 5. P10. L211. Please explain for what hypothesis test the t statistics are used in CRPS analyses. 6. Discussion. The next obvious question is where the difference between experts and non-experts come from. Is it because of their scientific knowledge, difference in information source or something else? I understand this is not the question the current study could address from the survey, but any discussion would be informative for readers. 7. Discussion. Does this study have any implication for the context outside of non-infectious diseases? In other words, to what extent this is related to the non-linear dynamic nature of infectious diseases? ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: Yes: Nikos I. Bosse Reviewer #2: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. 1 Apr 2021 Please find our response to reviewers' comments at the end of the full submitted PDF. Submitted filename: Response to reviewers.docx Click here for additional data file. 19 Apr 2021 How well did experts and laypeople forecast the size of the COVID-19 pandemic? PONE-D-21-04714R1 Dear Dr. Recchia, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Martial L Ndeffo Mbah, Ph.D Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: (No Response) Reviewer #2: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Thank you for addressing our comments in a very thorough manner. I feel the paper is much improved and am looking forward to seeing it published. Especially the discussion I found very good and helpful. I only have a few very minor notes about the paragraph on CRPS and the consensus distribution: - I think your explanation of the behaviour of the CRPS is very good and valuable. Maybe adding the word "precision" to the sentence "... serves as a measure of accuracy that can be applied to a consensus distribution constructed" would be a good idea, as the CRPS assesses sharpness subject to calibration (see Gneiting et al. 2007, Probabilistic forecasts, calibration and sharpness) - Explaining the CRPS in the context of one individual forecast is maybe slightly misleading, because that is exactly the scenario where the WIS should be used instead of the CRPS. If you have a limited set of quantiles, than the WIS is the better way to approximate CRPS than to assume an arbitrary distribution and calculate the CRPS based on that distribution. - In terms of the aggregate distribution I still find it hard to judge what really happens when you aggregate using assumed distributions instead of the WIS. I therefore still believe that using the WIS would be cleaner and simpler in terms of the statistical analysis. But I don't believe that this would change results in a meaningful way. Out of personal curiosity I played around with the data you provided and wrote some example code that gave striking results in the same direction you found. Please feel free to make use of this if you like, but also feel free to just ignore it. I think conclusions in the paper are fine without and that this shouldn't block publication. ```R library(data.table) install_github("epiforecasts/scoringutils") library(scoringutils) #read in data for experts and do same data manipulation to look at one of the targets # get a quantile column, rename as 'true_value' and 'prediction' data <- data.table::fread("uk_experts.csv") data <- data[, .(country_death_estimate, country_death_lower, country_death_upper)] data[, id := 1:nrow(data)] data <- melt(data, id.vars = "id") data[, true_value := 75346] data[variable == "country_death_estimate", quantile := 0.5] data[variable == "country_death_upper", quantile := 0.75] data[variable == "country_death_lower", quantile := 0.25] data[, variable := NULL] data[, model := "experts"] #read in data for high numeracy group and do same data manipulation to look at one of the targets data1 <- data.table::fread("uk_april_high_numeracy.csv") data1 <- data1[, .(country_death_estimate, country_death_lower, country_death_upper)] data1[, id := 1:nrow(data1)] data1 <- melt(data1, id.vars = "id") data1[, true_value := 75346] data1[variable == "country_death_estimate", quantile := 0.5] data1[variable == "country_death_upper", quantile := 0.75] data1[variable == "country_death_lower", quantile := 0.25] data1[, variable := NULL] data1[, model := "high_num"] data <- rbindlist(list(data, data1)) setnames(data, old = "value", new = "prediction") res <- eval_forecasts(data, summarise_by = "model") res wis_components(res) #could also be faceted for the different targets ``` Reviewer #2: The authors have adequately responded to my previous comments, and I recommend the current version of the manuscript will be published in the journal. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: Yes: Nikos I. Bosse Reviewer #2: No 28 Apr 2021 PONE-D-21-04714R1 How well did experts and laypeople forecast the size of the COVID-19 pandemic? Dear Dr. Recchia: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Martial L Ndeffo Mbah Academic Editor PLOS ONE
  17 in total

1.  Conditions for intuitive expertise: a failure to disagree.

Authors:  Daniel Kahneman; Gary Klein
Journal:  Am Psychol       Date:  2009-09

2.  Estimating the cost of regulating genome edited crops: expert judgment and overconfidence.

Authors:  Rim Lassoued; Peter W B Phillips; Stuart J Smyth; Hayley Hesseln
Journal:  GM Crops Food       Date:  2019-05-09       Impact factor: 3.074

3.  Predictive Mathematical Models of the COVID-19 Pandemic: Underlying Principles and Value of Projections.

Authors:  Nicholas P Jewell; Joseph A Lewnard; Britta L Jewell
Journal:  JAMA       Date:  2020-05-19       Impact factor: 56.272

4.  The role of complex thought in clinical prediction: social accountability and the need for cognition.

Authors:  J Ruscio
Journal:  J Consult Clin Psychol       Date:  2000-02

5.  Forecasting for COVID-19 has failed.

Authors:  John P A Ioannidis; Sally Cripps; Martin A Tanner
Journal:  Int J Forecast       Date:  2020-08-25

6.  A framework for evaluating epidemic forecasts.

Authors:  Farzaneh Sadat Tabataba; Prithwish Chakraborty; Naren Ramakrishnan; Srinivasan Venkatramanan; Jiangzhuo Chen; Bryan Lewis; Madhav Marathe
Journal:  BMC Infect Dis       Date:  2017-05-15       Impact factor: 3.090

Review 7.  Summary results of the 2014-2015 DARPA Chikungunya challenge.

Authors:  Sara Y Del Valle; Benjamin H McMahon; Jason Asher; Richard Hatchett; Joceline C Lega; Heidi E Brown; Mark E Leany; Yannis Pantazis; David J Roberts; Sean Moore; A Townsend Peterson; Luis E Escobar; Huijie Qiao; Nicholas W Hengartner; Harshini Mukundan
Journal:  BMC Infect Dis       Date:  2018-05-30       Impact factor: 3.090

8.  Short-term forecasting COVID-19 cumulative confirmed cases: Perspectives for Brazil.

Authors:  Matheus Henrique Dal Molin Ribeiro; Ramon Gomes da Silva; Viviana Cocco Mariani; Leandro Dos Santos Coelho
Journal:  Chaos Solitons Fractals       Date:  2020-05-01       Impact factor: 5.944

9.  Evaluating epidemic forecasts in an interval format.

Authors:  Johannes Bracher; Evan L Ray; Tilmann Gneiting; Nicholas G Reich
Journal:  PLoS Comput Biol       Date:  2021-02-12       Impact factor: 4.779

10.  Forecasting efforts from prior epidemics and COVID-19 predictions.

Authors:  Pranay Nadella; Akshay Swaminathan; S V Subramanian
Journal:  Eur J Epidemiol       Date:  2020-07-17       Impact factor: 8.082

View more
  6 in total

Review 1.  Complex correlates of Colombia's COVID-19 surge.

Authors:  James M Shultz; Ryan C Berg; Oscar Alberto Bernal Acevedo; Jovana Alexandra Ocampo Cañas; Victoria A Pinilla Escobar; Omar Muñoz; Zelde Espinel; José Miguel Uribe-Restrepo
Journal:  Lancet Reg Health Am       Date:  2021-09-10

2.  When Do We Need Massive Computations to Perform Detailed COVID-19 Simulations?

Authors:  Christopher B Lutz; Philippe J Giabbanelli
Journal:  Adv Theory Simul       Date:  2021-11-23

3.  Estimating and forecasting the burden and spread of Colombia's SARS-CoV2 first wave.

Authors:  Jaime Cascante-Vega; Juan Manuel Cordovez; Mauricio Santos-Vega
Journal:  Sci Rep       Date:  2022-08-09       Impact factor: 4.996

4.  An expert judgment model to predict early stages of the COVID-19 pandemic in the United States.

Authors:  Thomas McAndrew; Nicholas G Reich
Journal:  PLoS Comput Biol       Date:  2022-09-23       Impact factor: 4.779

5.  What happened to the predicted COVID-19-induced suicide epidemic, and why?

Authors:  Nick Glozier; Richard Morris; Stefanie Schurer
Journal:  Aust N Z J Psychiatry       Date:  2022-10-16       Impact factor: 5.598

6.  Comparing human and model-based forecasts of COVID-19 in Germany and Poland.

Authors:  Nikos I Bosse; Sam Abbott; Johannes Bracher; Habakuk Hain; Billy J Quilty; Mark Jit; Edwin van Leeuwen; Anne Cori; Sebastian Funk
Journal:  PLoS Comput Biol       Date:  2022-09-19       Impact factor: 4.779

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.