Literature DB >> 35965471

Bayesian emulation and history matching of JUNE.

I Vernon1,2, J Owen1,2, J Aylett-Bullock1,3, C Cuesta-Lazaro1,4, J Frawley1,5, A Quera-Bofarull1,4, A Sedgewick1,6, D Shi1,4, H Truong1,3, M Turner1,5, J Walker1,3, T Caulfield7, K Fong8,9, F Krauss1,3.   

Abstract

We analyze JUNE: a detailed model of COVID-19 transmission with high spatial and demographic resolution, developed as part of the RAMP initiative. JUNE requires substantial computational resources to evaluate, making model calibration and general uncertainty analysis extremely challenging. We describe and employ the uncertainty quantification approaches of Bayes linear emulation and history matching to mimic JUNE and to perform a global parameter search, hence identifying regions of parameter space that produce acceptable matches to observed data, and demonstrating the capability of such methods. This article is part of the theme issue 'Technical challenges of modelling real-life epidemics and examples of overcoming these'.

Entities:  

Keywords:  Bayes linear; calibration; disease models; emulation; history matching

Mesh:

Year:  2022        PMID: 35965471      PMCID: PMC9376712          DOI: 10.1098/rsta.2022.0039

Source DB:  PubMed          Journal:  Philos Trans A Math Phys Eng Sci        ISSN: 1364-503X            Impact factor:   4.019


Introduction: Bayesian emulation and uncertainty quantification

The COVID-19 pandemic disrupted healthcare systems and caused substantial fatalities around the globe. Various models have been developed to aid decision makers in the assessment of policy options, from simple analytic models up to complex agent-based models (ABMs). JUNE, introduced in [1], is of the latter type, and its high level of demographic and spatial resolution demands substantial computational resources to evaluate. A critical component in the uncertainty analysis, and subsequent use for decision support of a complex epidemiological model such as JUNE, is the process of model calibration: the matching of the model to observed data from the real system. This process can be extremely challenging, and in many cases, its intractability precludes the full exploitation of sophisticated models, which may otherwise contain nuanced insights into the system of interest. The problem of calibrating a complex and computationally demanding model is not unique to epidemiology and occurs in a wide range of scientific disciplines including cosmology, climate, systems biology, geology and energy systems [2-5]. To solve this problem, an area of Bayesian statistics arose, sometimes referred to as the uncertainty analyses of complex computer models, or to use its more recent (and slightly more general name): the area of uncertainty quantification (UQ) [5-7]. UQ provides a statistical methodology combining a large number of efficient techniques with a set of overarching principles that address how to analyze complex models rigorously, transparently and robustly, for use in scientific investigations, for making real-world predictions and for subsequent decision support. A core goal of this work is to demonstrate the capability of such methods for use with complex epidemiological models. A full analysis of the behaviour of models with a large number of input parameters and possibly several outputs, and their subsequent calibration, encounters the following three major issues: We summarize in the next section three UQ methods: (a) Bayes linear emulation, (b) linking models to reality and (c) Bayesian history matching, which address the aforementioned three problems. We then apply these UQ methods to the JUNE model in §3. For complex models, the evaluation time of the model is often so long that an exhaustive exploration of the model’s behaviour over its full input parameter space is infeasible. When comparing models to observed real-world data, an adequate statistical description of the link between model and reality, covering all major uncertainties and allowing for the rigorous use of an imperfect model, is required. When calibrating, the appropriate scientific goal should be to identify all locations in input parameter space that lead to acceptable fits between model and observed data, and not just to find the location of a single good match.

Bayesian emulation and history matching

Bayes linear emulation

Complex models typically have runtimes that can vary from seconds to days or even weeks, greatly inhibiting full model exploration, calibration, forecasting etc. Many of the techniques in UQ therefore revolve around the construction of Bayesian emulators: statistical constructs that mimic the scientific model in question, providing predictions of the model outputs with associated uncertainty, at as yet unevaluated input parameter settings [8]. The emulators provide insight into the model’s core structure and, unlike the models they mimic, are extremely fast to evaluate, typically being several orders of magnitude faster. Hence, they facilitate previously infeasible model exploration and global parameter searches. As an emulator makes predictions that have an associated (input dependent) uncertainty statement, they naturally fit within an overarching Bayesian uncertainty analysis, in which the impact of using an emulator instead of the model, can be understood and quantified. Emulators can be built for deterministic models, stochastic models, multilevel models (composed of models of increasing fidelity) and networks of models, providing a flexible and powerful set of tools to deal with a large class of scientific scenarios. Here, we outline the construction of Bayes Linear emulators, a robust form of emulator, based on a partial specification, which has been successfully employed in several settings [2,5]. We represent a general scientific model as the function . Here, is a vector composed of all the input parameters. For example, may represent an infectivity parameter, a social distancing parameter, etc. is the vector of all model outputs of interest, so, for example, may represent the number of people hospitalized in England on a particular day, may represent the number of deaths on that day, all as a function of the inputs . We denote the general component of as where the index will cycle through the full list of outputs of interest, for example, in the application to JUNE in §3, cycles through the set . We anticipate that, due to limited computational resources, we will only be able to evaluate the model at a finite (and possibly small) number of input parameter locations giving rise to model outputs , where , and ‘’ denotes the transpose. Therefore, at a new unevaluated input location, , even say for a deterministic (i.e. repeatable) model, we will still be uncertain about the output value of , as we will be for the majority of the input space. We take a subjective Bayesian view and treat the unknown as a random quantity and construct an emulator that represents our beliefs about possible reasonable forms that this function could take. A popular emulator form for each output is as follows [2]: where we have selected a subset of the inputs, , known as the active variables, , that are most influential for output . The first term on the right-hand side of equation (2.1) is a regression term, where are appropriately selected known deterministic functions of , a common choice being low-order polynomials, and are unknown scalar regression coefficients. The second term, , is a weakly second-order stationary process over , for which we only need to specify its second-order structure, choosing and utilizing an appropriate covariance function: a classic example suitable for smooth functions is the squared exponential: where and are the variance and correlation length of respectively, which may be specified a priori [2], or fitted using, e.g. MLE or MAP [4]. This simple covariance function may be enough, especially if the emulators regression term captures much of the model’s behaviour; however, if not, various extensions are available, e.g. individual correlation lengths for each input [9]. The third term, , is a white noise process uncorrelated with , , and itself such that with expectation zero, and . represents the effects of the remaining inactive inputs not included in the list of and formally facilitates a type of dimensional reduction [2]. The emulator form, as represented by equation (2.1), has various desirable features and exploits our beliefs about the general anticipated behaviour of physically realistic models. The regression term, , attempts to mimic the large-scale global behaviour of the function : often substantial in physical models. The second term, , the weakly stationary process, mimics the local behaviour of , again exploiting concepts of smoothness of either or attributes of (if, say, is stochastic). Such terms are highly versatile and can fit a large class of models; however, they require a sufficient density of runs to be suitably informed (regulated by the correlation length parameters ). In the literature, there is sometimes an over-reliance on similar Gaussian process style terms and a neglect of the regression terms, which may be unwise, as GPs of this form are typically capable of capturing the broad global behaviour, or the more complex local behaviour, but rarely both. We deliberately use the regression terms for the global structure and utilize the to capture the local behaviour. We can select the list of active inputs, , using various statistical techniques. For example, these could consist of classical linear model fitting criteria such as AIC or BIC, which have the benefit of speed and reasonable accuracy when applied to appropriate (nonlinear) sets of regression functions [2], or approaches such as automatic relevance determination [9], which can give increased accuracy provided the assumed form of the covariance function is suitable. In addition, we would also seek to incorporate expert knowledge of the model into the active input selection process, either by directly incorporating 'known' active inputs or by using a more nuanced Bayesian approach, of particular importance for expensive models. A list of active inputs for a particular physical output, , means that we have notably reduced the input dimensionality from to , which can result in large efficiency gains in subsequent calculations. The small remaining effect of the inactive inputs is not ignored, but is captured by the third term in equation (2.1), whose variance represents the added uncertainty induced by the dimensional reduction.

What to emulate

A major issue when emulating complex models is the choice of the set of attributes/outputs of the model to emulate. For example, often an objective function describing the mismatch between model and data has been emulated (e.g. a simple chi-squared metric, or a more complex likelihood function). However, despite being deceptively simple having just a single output, the objective function typically has a complex form, as it depends on the union of all active inputs and possesses numerous local maxima/minima [2], rendering this an inefficient strategy. Instead, we prefer to emulate the physical outputs of the model directly, as these tend to have (a) a smaller list of active inputs per output allowing a nuanced and sometimes substantial dimensional reduction tailored to each individual output, and (b) a simpler functional dependence on the input parameters that is often well represented by the regression terms in the emulator. Further choices are required when emulating stochastic models, where we can choose to emulate summaries of outputs of interest such as the mean, the variance or quantiles if required, possibly conditioning on key events such as epidemic take-off, and extend for example to covariance structures between groups of outputs if needed. In these cases, the role of is extended to also incorporate the uncertainties induced by using estimates from finite samples [10] or to employ full variance emulation as in ref. [11].

Designing batches of model evaluations

We begin by specifying the region of input space of interest, typically a -dimensional hypercube, and denote this . We then design a set of ‘space filling’ runs over , constructed to be well spread out, without any large holes. For example, we may use a maximin Latin hypercube design, an approximately orthogonal design, also desirable for emulator construction [12,13].

Updating the emulator

We then update our prior emulator structure given by equation (2.1) with the information from the set of model runs using our favourite statistical tools. Specifically, we would prefer a fully probabilistic Bayesian approach if we required full probability distributions on all emulated outputs, [14], and were we prepared to specify full joint probabilistic priors. However, in most cases, our preferred choice is to use Bayes Linear methods, a more tractable version of Bayesian statistics, which requires a far simpler prior specification and analysis [15,16]. It deals purely with expectations, variances and covariances of all uncertain quantities of interest, and uses the following Bayes linear update equations, derived from foundational arguments [16], to adjust our beliefs in the light of new data. When emulating the th output of a complex model, where we had performed an initial batch of runs giving a vector of model output values , we obtain the adjusted expectation, , and adjusted variance, , for at new input point using: and All quantities on the right-hand side of equations (2.4) and (2.5) can be calculated from equations (2.1) and (2.2) combined with prior specifications for , , , and . and provide a prediction for with associated uncertainty and are used directly in the implausibility measures used for the global parameter searches described in §2c. Note that multivariate versions of equations (2.4) and (2.5) are available. In addition, we may make certain pragmatic choices in the emulator construction process, for example, while we typically keep the regression coefficients uncertain, we may directly specify , and a priori, or use suitable plugin estimates as described in ref. [2]. We can test the emulators using a series of diagnostics, for example checking their prediction accuracy over a new batch of runs [17]. An example of a one-dimensional emulator is given in figure 1, cf. ref. [8] for an introduction and refs. [2,14,18] for details. The above Bayes linear emulation framework is fully implemented in the ‘hmer’ R package [19].
Figure 1

An emulator of a one-dimensional toy model, where , for the first wave/iteration, using just six runs (left panel), and for the second wave, using two additional runs (right panel). The emulator’s expectation and credible intervals are given by the blue and red lines, respectively, with the observed data that we wish to match to as the black horizontal line (with errors). The implausibility is represented by the coloured bar along the -axis, with dark blue implying , light blue and yellow (). (Online version in colour.)

An emulator of a one-dimensional toy model, where , for the first wave/iteration, using just six runs (left panel), and for the second wave, using two additional runs (right panel). The emulator’s expectation and credible intervals are given by the blue and red lines, respectively, with the observed data that we wish to match to as the black horizontal line (with errors). The implausibility is represented by the coloured bar along the -axis, with dark blue implying , light blue and yellow (). (Online version in colour.)

Assessing uncertainties: linking the model to the real world

Most epidemiology models are developed to help explain, understand and predict the behaviour of the corresponding real world system of interest, typically in terms of the progression through a population of an infectious disease. An essential part of determining whether such a model is adequate for this task is the comparison of the model with data formed from observations of the real system. However, this comparison involves several uncertainties that must be accounted for to provide a meaningful definition of an ‘acceptable’ match between a model run and the observed data. Hence, it is vital to define a clear statistical model describing the difference between the epidemiological model, , and the observed data denoted as the vector . While more complex statistical models are available [20], here, we describe a simple but powerful version that has been successfully used in a variety of scientific disciplines, for example climate, cosmology, oil reservoirs, epidemiology and systems biology [2-5,10,21]. The most familiar source of uncertainty is observational or experimental. We represent the features of interest of the real system as a vector of uncertain quantities, , which will be measured imperfectly involving a vector of errors , to give the vector of observations, , as follows: We represent the errors as additive here, but could use a more complex form if necessary. Depending on the scientific context, we then make judgements about the relationship between and , e.g. a common specification [2] is to judge the errors to be independent from , with expectation, and , a covariance matrix. Setting corresponds to the judgement that the observations were unbiased, and setting , that is a diagonal matrix, corresponds to uncorrelated observation errors, etc. A critical feature that we must incorporate is the difference between the epidemiological model, , of the system and the real system, , itself. We represent this difference between model and reality using a structural model discrepancy term. First, we note that even were we to evaluate the model, , at its best possible choice of input, , the output, , would still not be in agreement with the real epidemiological system value , due to the many simplifications and approximations inherent to the model; therefore, where is the structural model discrepancy: a vector of uncertain quantities that directly represents the difference between the model and the real system. Note that we are still treating , , and as vectors of random quantities. Now we have to make judgements about their relationships: a simple and popular specification [2,5] would be to judge that is independent of , and , with . In the case of a single output, we would then specify . However, for the full case of outputs, we may specify , a covariance matrix. may have intricate structure possessing non-zero covariances between components of , to capture the heavily correlated deficiencies of the model outputs. Various structures for of increasing complexity are available [2,5,14], along with methods for their specification [2,22]. Note that typically the form of is very different from . While the inclusion of the structural model discrepancy is unfamiliar to most modellers, it is now of standard practice in the UQ literature for analyzing complex but imperfect models [14,18,23,24]. It facilitates a richer analysis whereby we can incorporate our necessarily uncertain knowledge of the model’s deficiencies to improve our modelling of reality . Its inclusion can prevent over-fitting when calibrating and also reduces both bias and overconfidence when predicting. It is also vital when combining the results of several models.

Bayesian history matching

Due to their fast evaluation speed, emulators can be used in a variety of UQ calculations that would be otherwise infeasible. One of the most important is the problem of performing global parameter searches. Here, we outline a powerful iterative emulator-based global search method known as history matching (HM), which has been successfully employed in a variety of scientific disciplines [2,3,5,10]. HM is designed to answer the questions: Note the emphasis on finding all such acceptable matches: optimizing to find a single good fit is not adequate for assessing the impact of parametric uncertainty, nor for making predictions. Are there any input parameter settings that lead to acceptable matches between the model output and observed data? If so, what is the full set that contains all such input parameter settings? HM proceeds iteratively by ruling out regions of input parameter space that can be discarded from further investigation based on implausibility measures [5]. For an unexplored input parameter, , we can ask how far would the emulator’s expected value for the individual function output, , be from the corresponding observed value, , before we would deem it highly unlikely for to give an acceptable match were we to actually evaluate the function at . The implausibility measure, , captures this concept, and for an individual, output is given by the distance between emulator expectation and observed data, standardized by all relevant uncertainties, Here, is the emulator variance, is the variance of the model discrepancy and is the variance of the observational error, a direct consequence of equations (2.6) and (2.7). See also figure 1 (the -axis) for a depiction of . A large value of for a particular implies that we would be unlikely to obtain an acceptable match between and were we to run the model at . Hence, we can discard the input, , from the parameter search if , for some cutoff, , which is often chosen by appealing to Pukelsheim’s 3-sigma rule [25], a very general and powerful result, which states that for any continuous, unimodal distribution, 95% of its probability must lie within , regardless of asymmetry or skew, suggesting that a choice of may be reasonable [2]. This is the simplest univariate form, but we can combine implausibility measures from several outputs using say for some set , or employ more complex multivariate forms [2]. Before performing the th HM iteration, we define the current set of non-implausible input points as and the set of outputs that we considered for emulation in the previous wave as . We proceed as follows [4]: The history matching approach is powerful for several reasons: (a) while reducing the volume of the non-implausible region, we expect the function to become smoother, and hence to be more accurately approximated by the regression part of the emulator, . (b) At each new HM iteration, we will have a higher density of points and hence the second term, , in the emulator should be more effective, as it depends on proximity to the nearest runs. (c) In later iterations, the previously strongly dominant active inputs from early waves will have their effects curtailed, and hence, it will be easier to select additional active inputs, unnoticed before. (d) There may be several outputs that may be difficult to emulate in early iterations (perhaps because of their erratic behaviour in uninteresting parts of the input space) but simple to emulate in later waves once we have restricted the input space to a much smaller and more epidemiologically realistic region. See ref. [4] for further discussions comparing HM with Bayesian MCMC and ABC, ref. [26] for a direct comparison with ABC and the R package ‘hmer’ [19] for full implementation of the HM algorithm. We now apply these methods to the JUNE model. Design and evaluate a well chosen set of runs over the current non-implausible space , e.g. using a maximin Latin hypercube with rejection [2]. Combine these with any non-implausible runs surviving from previous waves. Check if there are new, informative outputs that can now be emulated accurately (that were difficult to emulate in previous waves) and add them to the previous set , to define . Use the runs to construct new, more accurate emulators defined only over the region for each output in . The implausibility measures , , are then recalculated over , using the new emulators. Cutoffs are imposed on the implausibility measures and this defines a new, smaller non-implausible volume , which should satisfy . Unless (a) the emulator variances for all outputs of interest are now small in comparison to the other sources of uncertainty due to the model discrepancy and observation errors, or (b) the entire input space has been deemed implausible, and return to step 1. If 6 (a) is true, generate as large a number as possible of acceptable runs from the final non-implausible volume , sampled depending on scientific goal.

Application of emulation and history matching to JUNE

The JUNE model

JUNE [1] is an ABM that describes the spread of an infectious disease through large synthetic populations. Originally designed to simulate the circulation of COVID-19 through the English population, JUNE has also been adapted to capture the populations of Cox’s Bazaar [27], a refugee camp in Bangladesh, and of Rhineland-Palatinate [28], one of Germany’s federal states. JUNE’s description of the epidemic spread rests on four areas: They are discussed in more detail below. the construction of a realistic synthetic population that reflects, as accurately as possible, the population demographic and their geographic distribution; the simulation of the population sociology, i.e. how the individuals behave: how they spend their time, whom they get into contact with and in which social environment; the parameterization of the infection, how it is transmitted from infected to susceptible individuals and impact it has on the health of infected individuals; the mitigation of spread and impact of the infection through pharmaceutical and non-pharmaceutical interventions (NPIs) such as social distancing and vaccinations, respectively.

Population

JUNE builds its synthetic population based on real or parameterized census data—in the case relevant for this contribution, JUNE constructs the about 55 million residents of England based on the 2011 census data accessible through the NOMIS database provided by the ONS. The data are organized hierarchically, with Output Areas (OAs) the smallest relevant unit, comprising on average about 250 residents with relatively similar socio-economic characteristics. The OAs have a specified geographic location, and their data contain information about age, sex and ethnicity of the area’s residents [29-31] and the composition of the households they live in [32], in about 20 categories.[1] JUNE uses national averages to correlate age, sex and ethnicity of individuals, which are presented as uncorrelated distributions in the data. In a similar way, information such as the national distributions of age differences of partners [33], and of parents and their children [34], are used to assign the individuals to their households. As additional static properties of the population, JUNE assigns school-aged children to the nearest age-appropriate school; information about school locations and the age ranges for their students is taken from ref. [35]. Within the schools, the students are grouped into class units of 20–30 individuals and have teachers assigned to them. In a similar way, universities are filled with students—the young adults—and they are grouped into year groups of about 200 students. The OAs are part of Middle Super Output Areas (MSOAs) with about 12 500 residents and 50 OAs constituting one MSOA. The census data provide information about the sectors of companies within MSOAs and about the distribution of the working population over these sectors, using the Standard Industrial Classification (SIC) scheme [36]. The parameterization of company sizes with national sector-dependent averages allows JUNE to construct an origin-destination matrix for the employees at the level of MSOAs [37]. Information concerning the commuting habits of individuals contained in the census data [38] underpins the construction of simplified virtual public transport networks within JUNE.

Interactions

Having defined the static properties of the synthetic population—where people live, work and study—their daily lives outside work and education are filled with various activities. These activities include shopping, visiting friends and relatives in their homes, frequenting pubs and restaurants, going to the gym or cinema, to name a few. In the absence of any of these leisure activities, people are supposed to stay at home. Surveys performed, e.g. by the Office for National Statistics [39], define the average proportion of time spent with various activities, in dependence on age and sex. These averages are translated into a probabilistic treatment thereby creating a highly flexible and varied daily schedule for JUNE’s virtual individuals. These schedules are supplemented with contact matrices from PolyMod [40] and the BBC Pandemic Project [41], which indicate the average number of daily contacts—communication or physical—of individuals of age with individuals of age in different social settings , for example home (), school () and work (). As the contact numbers are presented as population averages, suitable for their deployment in compartment models, they need to be renormalized for the socially more granular IBMs,[2] resulting in the renormalized overall contact matrices and the corresponding fraction of physical contacts, , where . While this introduces some uncertainty into the modelling of social interactions, the interplay of the synthetic population model with the contact matrices provides a welcome closure test for the self-consistency of the overall model. For the purpose of fitting to data and the quantification of uncertainties in the model, we assume that the construction of the synthetic population and its interactions is well understood and robustly and well parameterized as it is driven by data of relatively high quality.

Infection

The description of the infection consists of two separate parts. First, the transmission from an infected person to a susceptible person needs to be simulated. In JUNE, as in many other models, this is described as a probabilistic process. The infection probability for a susceptible person with susceptibility during a time interval from to , spent with a group of individuals in social context is given as follows: In the aforementioned equation, denotes the time-dependent infectiousness of the infected individual in group . In JUNE, it follows a profile given by with , and is the time of infection of the individual, is the incubation period and is the gamma function. is sampled from a normal distribution. The maximal or peak value of infectiousness for individual is sampled from a log-normal distribution with median and shape parameter , which allows for a long but small tail of highly infectious individuals, which can be connected to super-spreader events. The in equation (3.1) is the contact intensity between and , where are the social location-dependent baseline intensities, is the number of individuals in the group setting, normalizing the contact number and parameterizes the relative increase in infection probability for the proportion of physical contacts . These parameters, the social-environment dependent and the universal , cannot be derived from first principles and must be obtained from fits to available data; they constitute a significant portion of the parameter space in the model and, correspondingly, a significant source of uncertainty. Once an individual is infected, it takes some time—the incubation period—before they can infect others and some additional time before the onset of symptoms. A large range of input data has been used to derive various symptom trajectories for infected individuals, which in the case of high-income western countries in the global North depends mainly on their age and sex.[3] In the original formulation of the JUNE model, significant efforts have gone into the quantification of probabilities for different health outcomes in the population, with some emphasis to also capture the health impact of COVID-19 on the highly vulnerable care home residents; we refer the reader to ref. [1] for more details. Here, it should suffice to state that in JUNE asymptomatic and symptomatic trajectories with varying severity have been identified, the latter ranging from mild, flu-like symptoms over admission to regular or intensive-care wards to death in hospital or at residence. Although there are some uncertainties related to this treatment, we usually do not consider them and treat the health outcomes as fixed by data. We seed initial infections based on the number of fatalities 2–3 weeks afterwards, by using the infection-fatality rates obtained from data and encoded in JUNE. The parameter is an additional factor that modifies the resulting number of initial infections.

Interventions

Since the beginning of the COVID-19 epidemic, the UK government—like many other governments around the world—has employed a wide range of mitigation strategies. At the beginning of the pandemic, these interventions were mainly non-pharmaceutical, and these NPIs ranged from relatively simple strategies at the level of individuals, such as mask wearing and other social distancing measures, to more involved and global strategies such as partial or complete lockdowns, involving school closures and the furloughing of parts of the work force. In JUNE, these measures can easily be modelled: social distancing measures and mask wearing can be described by modifying the ’s in the corresponding social settings by a factor, , capturing the reduced, but non-zero, transmission probability, while the closure of schools or furloughing of the work force is easily described, based on data [42-45], by keeping the impacted population at home instead of sending them to schools or work. For a more detailed description of the translation of NPIs to the JUNE simulation, we refer the reader to ref. [1].

Inputs, outputs and initial emulation

Our primary goal is to test if the JUNE model can produce acceptable matches to observed data at the national and regional level, from the first wave of the COVID-19 pandemic and the subsequent summer period. We wish to identify the region of parameter space, , leading to such acceptable fits, if it exists. We identify a large set of input parameters, , of interest to search over, primarily composed of interaction intensity parameters at the group level, seeding and quarantine compliance parameters, and social distancing parameters (see [1] for details), and specify ranges for each, given in table 1, which define the initial search space, . These ranges were chosen to be conservative, informed in part by earlier exploratory runs while also respecting the role each parameter plays in the model, including the time period over which they operate (see [1] for details). A typical full England run of JUNE would take approximately 10 hours to complete on 64 cores (Intel Xeon Skylake) and 128 GB of memory. This substantial computational expense combined with a relatively high-dimensional input parameter space makes a global parameter search extremely challenging and necessitates the use of emulation. While there are several types of data available for the early pandemic, many of these had questions regarding reliability. For example, case data were substantially affected by the limited and evolving availability of COVID-19 tests, while hospital admission data were, especially during the peak of the first wave, collected with understandably varying levels of diligence across trusts. While it would in principle be possible to incorporate such data sets using a detailed statistical model for the measurement errors, , in equation (2.6) that incorporated under-counting, we instead focus on hospital and total death data [46-49], which although still uncertain due to the precise definition of death with COVID-19, suffers from far fewer issues.
Table 1

The input parameters explored in the global parameter search, their type and their ranges that define the search region .

input parameter (xi)typerange
βpublocation-dependent contact intensity[0.02,0.6]
βgrocery[0.02,0.6]
βcinema[0.02,0.6]
βuniversity[0.02,0.6]
βcitytransport[0.08,0.77]
βintercitytransport[0.08,1.2]
βhospital[0.08,1.2]
βcarehome[0.08,1.2]
βcompany[0.08,1.2]
βschool[0.08,1.2]
βhousehold[0.08,1.2]
βcarevisits[0.1,10]
βhouseholdvisits[0.1,10]
αphysicalphysical contact factor[1.8,3]
αseedstrengthmodifies initial/seeding infections[0.1,2]
Mquarantinehouseholdcompliancequarantine compliance[0.034,0.26]
Msocialdistancingβfactorsocial distance (1 week prior to lockdown)[0.65,0.95]
Msd3randomfactorallenhanced social distance (full lockdown)[0.1,0.5]
Msd4randomfactorallsocial distance (post lockdown, non-leisure)[0.25,1]
Msd4randomfactorleisuresocial distance (post lockdown, leisure)[0.25,1]
The input parameters explored in the global parameter search, their type and their ranges that define the search region . We define the primary JUNE outputs of interest to be the hospital deaths and total deaths from 19 March to the end of August 2020, for England and its seven regions: East of England, London, Midlands, North East and Yorkshire, North West, South East and South West. We choose a subset of dates , shown as the vertical dashed lines in figure 4, to emulate. The observed data and JUNE output is noisy, so we smooth them both using a standard kernel smoother (Gaussian kernel, bandwidth 7 days) as we wish to compare the underlying trends, and define the smoothed versions to be the target observed data points (shown in figure 2), and the primary JUNE outputs, . Therefore, index cycles through elements of the set , where labels hospital or total deaths, labels each of the seven regions of England or England itself and labels the time points of interest, given as the dashed lines in figure 4. We specify conservative observation error and model discrepancy variances and for each output as described in ref. [1], by decomposing each into multiplicative and additive components to represent possible systematic biases, in addition to a scaled component for the observation error only, to model the noisy count process.
Figure 4

The JUNE output for total daily deaths in England in 2020, for several iterations of the HM process. The smoothed and noisy data, along with the combined uncertainties due to and , are shown in black. (Online version in colour.)

Figure 2

Daily deaths in hospital wards and ICU in 2020, by region. The smoothed version used in the HM is also shown. (Online version in colour.)

Daily deaths in hospital wards and ICU in 2020, by region. The smoothed version used in the HM is also shown. (Online version in colour.) We design and evaluate a first iteration/wave of 150 runs over the input space, , using a maximin Latin hypercube design. The outputs of these runs are shown as the purple lines in figure 4. We construct emulators for each output, , as detailed in §2a (using full quadratic regression terms selected using BIC, and MAP estimates for [2]). The emulators provide insight into the behaviour of the JUNE model. For example, we can examine the coefficients of the linear terms for the inputs featuring in equation (2.1), to gain insight into the effect each input has on each individual output. Estimates of these are shown in figure 3 for the total deaths in England outputs, where index therefore cycles through just the various time points: with each time point, labelled on the -axis, giving rise to a single vertical strip in the plot corresponding to a single emulator (note that a finer temporal resolution is used here for added detail, while far fewer time points are used in the HM). Conversely, labels the active input in question as given on the -axis. Here, red/blue represents positive/negative dependencies , respectively, standardized as proportions of the largest coefficient of that output. We see strong anticipated contributions from , and in the first wave of the pandemic from March to May, and more modest effects from and throughout the summer period. The sensitivities of and change to negative (blue) by May, as in many of these uncalibrated runs, herd immunity has been reached, and hence increasing will decrease deaths (as they will be brought forward in time). Note that the parameters and are not included in figure 3 as they were only implemented prior to the second iteration of runs, but were included in the subsequent full history match by suitable inflation of the iteration 1 emulator uncertainties [20]. More insight can be gained from full emulator sensitivity analysis [50].
Figure 3

Estimates for the coefficients of the linear terms that are found to feature in the emulators for total deaths in England for the first iteration/wave of runs, where labels the time point (-axis) and labels the inputs (-axis). Red/blue represents positive/negative dependencies of on that input, respectively, standardized as proportions of the largest coefficient for that time point. A finer temporal resolution is used here for added clarity. Note that this plot shows the time-dependent sensitivity of the model to the inputs, but that the actual inputs do not vary over time. (Online version in colour.)

Estimates for the coefficients of the linear terms that are found to feature in the emulators for total deaths in England for the first iteration/wave of runs, where labels the time point (-axis) and labels the inputs (-axis). Red/blue represents positive/negative dependencies of on that input, respectively, standardized as proportions of the largest coefficient for that time point. A finer temporal resolution is used here for added clarity. Note that this plot shows the time-dependent sensitivity of the model to the inputs, but that the actual inputs do not vary over time. (Online version in colour.) The JUNE output for total daily deaths in England in 2020, for several iterations of the HM process. The smoothed and noisy data, along with the combined uncertainties due to and , are shown in black. (Online version in colour.)

Iterative history matching

We now employ the history matching framework from §2c, iteratively removing parameter space based on current implausibility measures, and performing batches/iterations of further runs. Initially, in iterations 1 and 2, only the hospital and total deaths for England up to the end of May 2020 were included in the HM, to rule out the more exotic regions of parameter space, while for iterations 3–5, all the seven regions of England were also included and the time period extended to the end of August 2020. Figure 4 shows the outputs from iterations 1, 3 and 5 for total deaths in England as the purple, yellow and red lines, respectively. As the iterations proceed, the emulators become more accurate, we learn more about the global parameter space, and hence, the runs approach the observed data, yielding reasonable matches across the first COVID-19 wave. By iteration 5, the majority of emulators attained the accuracy required for the stopping criteria in the HM algorithm. The region of 20-dimensional parameter space deemed non-implausible at iteration 5 is shown in figure 5 as a collection of two-dimensional optical depth plots, which simply show the depth in the remaining 18 dimensions of the non-implausible region (see [4]). The optical depth is defined for each point in the two-dimensional space shown in each individual plot panel as follows: where denotes the 18-dimensional volume of the remaining space. can therefore show where large or small amounts of non-implausible points can be found, conditioned on , providing further insight into the structure of . Figure 5 gives insights into the constraints imposed on the parameters by the death data and corresponding uncertainty specification. For example, we see that we learn a lot about certain influential parameters such as , which are fairly well constrained, while others such as can take a wider range of values. Provisional investigations suggest we can further constrain by adding deaths in care home settings to the calibration outputs. We also see interesting relationships between pairs of parameters, e.g. the reciprocal relations between vs. , suggesting one or other can be high, but not both. We see similar relations between vs. . However, one should be aware that the actual constraints imposed are higher dimensional in nature and cannot be fully represented by such two-dimensional plots, but that they can be explored further e.g. by examining the eigenstructure of , as done in ref. [51]. Note that in using HM in this way, we do not seek to probabilize the non-implausible region as in a full Bayesian calibration, but we could go on to do this (e.g. by routing the emulators through an MCMC algorithm) if desired, but the additional information gained may be in part an artefact of the particular additional distributional choices that such an analysis requires, which may impact robustness and predictive accuracy.
Figure 5

The optical depth of various two-dimensional projections of the full 20-dimensional non-implausible region found after the 5th iteration. The 12 most constrained inputs are show, labelled on the diagonal (the remaining eight inputs were relatively unconstrained). The colour scales are standardized and linear in depth, with yellow showing maximum depth for that projection and purple/black showing minimum/zero depth. This region corresponds to the red runs in figure 4. (Online version in colour.)

The optical depth of various two-dimensional projections of the full 20-dimensional non-implausible region found after the 5th iteration. The 12 most constrained inputs are show, labelled on the diagonal (the remaining eight inputs were relatively unconstrained). The colour scales are standardized and linear in depth, with yellow showing maximum depth for that projection and purple/black showing minimum/zero depth. This region corresponds to the red runs in figure 4. (Online version in colour.) While performing a full global exploration of the input parameter space is of course preferable, it is sometimes useful to perform a fast ‘look-ahead’ stage to check if such an expensive model is capable of fitting the next period of observed data, or whether model improvements are required. Figure 4 also shows the results of such an exercise, where we took eight runs with acceptable matches to the death data up to the end of August, and performed small 30-point five-dimensional Latin hypercube designs for each of the eight cases up until December 2020, now varying only five additional parameters relevant to the second COVID-19 wave (social distancing for schools, leisure, and non-leisure activities, November lockdown and B.1.1.7 variant infectiousness), the output of which is given by the green lines. One iteration of HM was performed to reduce the five-dimensional parameter space in each case, and a new set of runs designed, which are shown as blue lines in figure 4. We see reasonable matches to the first part of the second COVID-19 wave, with perhaps a late take-off in early September, and a partial overshoot in November to December, suggesting that JUNE may well provide acceptable matches after a full HM. To give more detail, figure 6 shows a single unsmoothed run (red lines), from this final batch, but now for hospital deaths and total deaths for England and all seven regions, and shows the sort of quality of matches we are seeing so far. The black points give the (unsmoothed) death data and the combined uncertainties due to and shown as the blue lines. The fact that JUNE matches several regions simultaneously, at least over the first wave, without resorting to any region specific parameters, suggests that geographical variations in the relative importance in different types of interaction drove/affected the different epidemic curves in those regions. Further, more detailed investigation to confirm this is of course required. We leave the extension to 2021 and beyond to the future work, as this requires the complex behavioural and partial restrictions (on travel, visiting relatives, etc.) imposed over the December 2020 Christmas period (the ‘cancelled Christmas’) and the January to April 2021 lockdown and subsequent staggered release to be implemented and tested, possibly requiring additional time-dependent parameters, which is the ongoing work (although we note that vaccines and multiple variants have already been implemented in JUNE). We also note that the process of using complex models combined with emulators and appropriate uncertainties to make realistic predictions over such periods is a substantive UQ topic in its own right, which deserves separate treatment [52].
Figure 6

A single JUNE run (red lines), from the second exploratory iteration (i.e. one of the blue lines in figure 4). The panels show hospital deaths (rows 1 and 2, viewed in landscape) and total deaths (rows 3 and 4, viewed in landscape) for England and the seven regions, as given in the plot titles. The black points give the (unsmoothed) death data, and the combined uncertainties due to and are shown as the blue lines. (Online version in colour.)

A single JUNE run (red lines), from the second exploratory iteration (i.e. one of the blue lines in figure 4). The panels show hospital deaths (rows 1 and 2, viewed in landscape) and total deaths (rows 3 and 4, viewed in landscape) for England and the seven regions, as given in the plot titles. The black points give the (unsmoothed) death data, and the combined uncertainties due to and are shown as the blue lines. (Online version in colour.)

Discussion

Models such as JUNE, with its high level of demographic and spatial granularity, may become important tools to aid local and national decision makers. However, to fully exploit the nuances of such complex and expensive ABMs, efficient and comprehensive calibration methods are required. We demonstrated the emulation of JUNE, providing insight into the model structure, and employed HM to identify the region of parameter space yielding reasonable matches to national and regional level hospital and total death data for the first COVID-19 wave. Such techniques form an essential tool for the future use of complex epidemiological ABMs, expanding our capabilities to combine detailed models with rigorous UQ. The ability to perform global exploration of the parameter spaces of expensive models of this form and to embed this within a broader UQ framework is vital for making predictions with realistic uncertainty statements and hence vital for subsequent decision support.

Outlook/future directions

Our work represents an important step towards the full exploitation of highly granular and detailed ABMs in health settings and elsewhere, harnessing the full depth of their simulations in providing high-quality understanding of critical dynamics and robust quantitative projections for improved decision support. The next steps in this project are to include further outputs of interest within the HM for JUNE, (hospitalizations, case rates, age categories, etc.) and to examine smaller geographic regions, in which the stochasticity of JUNE will become more pronounced, compared to the national/regional level where it is somewhat subdominant. This will require more sophisticated emulator strategies [11], and if we are interested in detailed spatial predictions, will require the updating of the JUNE state vector using UQ style data-augmentation techniques [53]. Beyond this, these UQ methods are currently being incorporated wherever JUNE is being employed e.g. by the UN for Cox’s Bazaar [27], a refugee camp in Bangladesh, and for Rhineland-Palatinate [28], one of Germany’s federal states. In addition, we plan to use the model to investigate in more detail social imbalances in COVID-19 attack rates and infection-fatality ratios, which are relatively easy to trace in a model such as JUNE. Supplementing the model with the elaborate UQ techniques will allow us to identify, in more detail and with increased certainty, important correlations between socio-economic markers of the population and the infection dynamics and outcomes.
  8 in total

1.  Understanding hormonal crosstalk in Arabidopsis root development via emulation and history matching.

Authors:  Samuel E Jackson; Ian Vernon; Junli Liu; Keith Lindsey
Journal:  Stat Appl Genet Mol Biol       Date:  2020-07-13

2.  Contagion! The BBC Four Pandemic - The model behind the documentary.

Authors:  Petra Klepac; Stephen Kissler; Julia Gog
Journal:  Epidemics       Date:  2018-03-22       Impact factor: 4.396

3.  June: open-source individual-based epidemiology simulation.

Authors:  Joseph Aylett-Bullock; Carolina Cuesta-Lazaro; Arnau Quera-Bofarull; Miguel Icaza-Lizaola; Aidan Sedgewick; Henry Truong; Aoife Curran; Edward Elliott; Tristan Caulfield; Kevin Fong; Ian Vernon; Julian Williams; Richard Bower; Frank Krauss
Journal:  R Soc Open Sci       Date:  2021-07-07       Impact factor: 2.963

4.  Bayesian history matching of complex infectious disease models using emulation: a tutorial and a case study on HIV in Uganda.

Authors:  Ioannis Andrianakis; Ian R Vernon; Nicky McCreesh; Trevelyan J McKinley; Jeremy E Oakley; Rebecca N Nsubuga; Michael Goldstein; Richard G White
Journal:  PLoS Comput Biol       Date:  2015-01-08       Impact factor: 4.475

5.  History matching of a complex epidemiological model of human immunodeficiency virus transmission by using variance emulation.

Authors:  I Andrianakis; I Vernon; N McCreesh; T J McKinley; J E Oakley; R N Nsubuga; M Goldstein; R G White
Journal:  J R Stat Soc Ser C Appl Stat       Date:  2016-11-24       Impact factor: 1.864

6.  Bayesian uncertainty analysis for complex systems biology models: emulation, global parameter searches and evaluation of gene functions.

Authors:  Ian Vernon; Junli Liu; Michael Goldstein; James Rowe; Jen Topping; Keith Lindsey
Journal:  BMC Syst Biol       Date:  2018-01-02

7.  Social contacts and mixing patterns relevant to the spread of infectious diseases.

Authors:  Joël Mossong; Niel Hens; Mark Jit; Philippe Beutels; Kari Auranen; Rafael Mikolajczyk; Marco Massari; Stefania Salmaso; Gianpaolo Scalia Tomba; Jacco Wallinga; Janneke Heijne; Malgorzata Sadkowska-Todys; Magdalena Rosinska; W John Edmunds
Journal:  PLoS Med       Date:  2008-03-25       Impact factor: 11.069

  8 in total
  1 in total

1.  Visualization for epidemiological modelling: challenges, solutions, reflections and recommendations.

Authors:  Jason Dykes; Alfie Abdul-Rahman; Daniel Archambault; Benjamin Bach; Rita Borgo; Min Chen; Jessica Enright; Hui Fang; Elif E Firat; Euan Freeman; Tuna Gönen; Claire Harris; Radu Jianu; Nigel W John; Saiful Khan; Andrew Lahiff; Robert S Laramee; Louise Matthews; Sibylle Mohr; Phong H Nguyen; Alma A M Rahat; Richard Reeve; Panagiotis D Ritsos; Jonathan C Roberts; Aidan Slingsby; Ben Swallow; Thomas Torsney-Weir; Cagatay Turkay; Robert Turner; Franck P Vidal; Qiru Wang; Jo Wood; Kai Xu
Journal:  Philos Trans A Math Phys Eng Sci       Date:  2022-08-15       Impact factor: 4.019

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.