Literature DB >> 27293793

Fast methods for training Gaussian processes on large datasets.

C J Moore1, A J K Chua1, C P L Berry2, J R Gair3.   

Abstract

Gaussian process regression (GPR) is a non-parametric Bayesian technique for interpolating or fitting data. The main barrier to further uptake of this powerful tool rests in the computational costs associated with the matrices which arise when dealing with large datasets. Here, we derive some simple results which we have found useful for speeding up the learning stage in the GPR algorithm, and especially for performing Bayesian model comparison between different covariance functions. We apply our techniques to both synthetic and real data and quantify the speed-up relative to using nested sampling to numerically evaluate model evidences.

Entities:  

Keywords:  Gaussian processes; data analysis; inference; regression

Year:  2016        PMID: 27293793      PMCID: PMC4892455          DOI: 10.1098/rsos.160125

Source DB:  PubMed          Journal:  R Soc Open Sci        ISSN: 2054-5703            Impact factor:   2.963


Introduction

A wide range of commonly occurring inference problems can be fruitfully tackled using Bayesian methods. A particularly common inference problem is that of regression; determining the relationship of a control variable x to an output variable y given a set of measurements of {y} at points {x}. The solution requires a model y=f(x), which allows us to predict the value of y at an untested value of x. From a Bayesian standpoint, this can be achieved using Gaussian processes (GPs): a GP is a collection of random variables, of which any finite subset have a joint Gaussian probability distribution [1]. Gaussian process regression (GPR) is a powerful mathematical technique for performing non-parametric regression in a Bayesian framework [1-5]. The key assumption underpinning the method is that the observed dataset being interpolated is a realization of a GP with a particular covariance function. This assumption presents us with a challenge: how do we choose the covariance function which gives the best interpolant? The process of choosing the covariance function is known as learning, or training of the GP. In this training process, it is necessary to compute the inverse of the covariance matrix (the matrix formed by evaluating the covariance function pairwise between all n observed points). The time taken to evaluate the inverse of the covariance matrix scales as [1], where n is the number of points being interpolated; this has typically restricted the application of GPR to smaller problems (), although work has been done on extending its applicability to larger datasets [6-9]. In this paper, we present two techniques that speed up the training stage of the GPR algorithm. The first aims to reduce the dimensionality of the problem, and hence speed up the learning of the hyperparameters for a single covariance function. This does not change the fact that the cost of this process is ; instead it simply reduces the constant in this scaling. The second aims to enable fast Bayesian model comparison between different covariance functions while also incorporating the benefits of the first technique. We consider maximizing the hyperlikelihood: the conditional probability of the data given a particular set of hyperparameters used to specify the covariance function.[1] We provide an expression for the Hessian matrix of the hyperlikelihood surface and show how this can be used as a valuable tool for comparing the performance of two different covariance functions. We also present modified expressions for the hyperlikelihood, its gradient and its Hessian matrix, which have all been analytically maximized and marginalized over a single-scale hyperparameter. This analytic maximization or marginalization reduces the dimensionality of the subsequent optimization problem and hence further speeds up the training and comparison of GPs. These techniques are useful when attempting to rapidly fit large, irregularly sampled datasets with a variety of covariance function models. The authors have previously made use of these techniques in exploring the correlation structure of the differences between complicated waveform models in the field of gravitational-wave astronomy [10,11]; this was done so that the effect of different models on the parameter inferences could be marginalized over. There, the behaviour of the data was largely unknown a priori and it was necessary to quantitatively compare a wide range of different covariance functions. Work in this area with larger datasets is ongoing. In §2, we review the GPR method and discuss methods of efficiently determining a covariance function. In §2.1, we present our expression for the Hessian of the hyperlikelihood along with a discussion of how it can be used for model comparison, and in §2.2, we show how the training of the GP can be accelerated by analytically maximizing or marginalizing the hyperlikelihood over a single-scale parameter. In §3, we apply these methods to both synthetic and real datasets, and compare the computational cost to that of a full numerical evaluation of the Bayesian model evidences. Finally, a brief discussion and concluding remarks are given in §4.

Gaussian process regression and training

The technique of GPR is a method for interpolating (or extrapolating) the data contained in a training set . The vector ={x | i=1,2,…,n} is called the input vector and the output vector is given by y=f(x) for some unknown function f. The method works by assuming that the data have been drawn from an underlying GP with specified mean μ(x) (usually assumed to be zero) and covariance function k(x,x′). There is freedom in specifying the covariance function; common choices, such as the squared exponential and Matérn function, include a number m of free hyperparameters ={θ | i=1,2,…,m} that control the properties of the GP, i.e. k(x,x′)=k(x,x′;). The predictive power of the method comes from computing the conditional probability of the function taking a given value at some new (n+1)th input point x*, given the observed values in and the values of the hyperparameters . This predictive probability distribution for the function at the new point is a Gaussian with mean and variance [σ(x*)]2 [1], and where we have defined the scalar, vector and matrix shorthand Since the posterior distribution for (2.2) relies upon the form of the covariance, GPR cannot be used to make definite predictions until we have fixed a method for dealing with the unknown hyperparameters . Ideally, we place a prior probability distribution on and make predictions by evaluating the integral where we have used Bayes’ theorem to obtain the second equality. We have introduced the hyperlikelihood given by which encodes the probability that the observed (training) data were drawn from a GP with covariance function k. The integral (2.4) is almost always analytically intractable and prohibitively expensive to evaluate numerically. A common approximate approach is to use the most probable values of the hyperparameters , which maximize [12-14]. Assuming the prior distribution is sufficiently flat (or uninformative) over the region of interest, this is equivalent to maximizing the hyperlikelihood P( | ,). Under this approximation, the predictive distribution becomes which is simply the Gaussian in (2.2) with mean and variance evaluated at . Implementing the above procedure requires numerically maximizing the hyperlikelihood in (2.5). This can be computationally expensive; in §2.1 and §2.2, we present methods for reducing the cost of maximizing the hyperlikelihood.

Using the gradient and Hessian

The maximization process may be accelerated if the gradient of the hyperlikelihood is known and a gradient-based algorithm, such as a conjugate gradient method [13,15], can be used. The gradient of the logarithm of the hyperlikelihood is given by [1] This can be shown by differentiating (2.5) and making use of the standard results The gradient in (2.7) is useful because the rate-determining step in computing the hyperlikelihood is computing the inverse matrix K−1 (usually achieved through a Cholesky decomposition in practice), which is an operation. All other steps in (2.5) scale as or less.[2] Once the inverse has been calculated, the gradient in (2.7) may also be evaluated in ; so in evaluating the hyperlikelihood for a large training set we can also get the gradient for negligible extra cost. The procedure outlined above can be performed for multiple covariance functions, each yielding a different GP interpolant. It is, therefore, necessary to have a method of comparing the performance of different interpolants to decide which to use. One way to achieve this is to evaluate the (hyperprior-weighted) volume under the hyperlikelihood surface, the hyperevidence, and use this as a figure of merit for the performance. Evaluating this integral is prohibitive, so an approximation is to calculate the Hessian matrix of the surface at the peak (the position and value of which have already been found) and to analytically integrate the resulting Gaussian. This procedure assumes flat (or slowly varying) hyperpriors in the vicinity of the peak, but this has already been assumed in going from (2.4) to (2.6). Differentiating the gradient in (2.7), again making use of the results in (2.8), and evaluating the derivatives at the position of peak hyperlikelihood, , gives the Hessian, This expression has the same advantages as the expression for the gradient; as the inverse of the covariance matrix has already been computed, the Hessian may be evaluated at negligible extra cost. The hyperlikelihood surface may therefore be approximated by the Gaussian [12,16] We seek the hyperevidence, which is given by the following integral of the hyperposterior, where we have specified a prior Π() on the hyperparameters; Assuming the hyperposterior is a sufficiently well-peaked distribution, with peak at position , the hyperevidence may be written using the Laplace approximation [2] as It is always possible to change the hyperparametrization so that the prior is flat in which case the hyperposterior is proportional to the hyperlikelihood.[3] If such a hyperparametrization has been chosen then (where V is the hyperprior volume, or range of integration), =0 and ; therefore This expression is now invariant under further changes to the hyperparameter specification which preserve the property that the prior is constant. We use hyperparametrizations with flat hyperpriors as this choice uniquely specifies the approximation in equation (2.13); although there remains the possibility that another hyperparametrization exists in which the posterior is better approximated as a Gaussian. For two covariance functions, k1 and k2, the odds ratio may be defined as the ratio of the value of (2.13) evaluated with k1 to the value evaluated using k2, and this may be used to discriminate among competing models. The hyperprior volume V in (2.13) acts as an Occam factor, penalizing models with greater complexity [2]. Once suitable prior volumes have been fixed, the Hessian approximation to the hyperevidence is a computationally inexpensive means of comparing covariance functions. The Hessian may also be used to provide error estimates for the hyperparameters; from (2.10) it can be seen that the inverse of the Hessian is the covariance matrix of the maximum hyperlikelihood estimator of the hyperparameters.

Partial analytic maximization

In general, covariance functions can be arbitrarily complicated, with large numbers of hyperparameters. Inevitably, simple covariance functions are the most prevalent in the literature. If there are a small number of hyperparameters, then even reducing the number of hyperparameters by one can have a great impact on the length of time taken to maximize the hyperlikelihood. In this section, we show how the hyperlikelihood for any covariance function, regardless of complexity, can be analytically maximized over an overall scale parameter, thereby reducing the number of remaining hyperparameters. We also generalize the expressions for the gradient and the Hessian found in §2.1 to this case. Consider the following transformation of the covariance, ; substituting this into the expression for the hyperlikelihood gives, This function always has a unique maximum with respect to variations in at the position at this point, the hyperlikelihood takes the value Equation (2.16) is to be considered as a function of the remaining m−1 hyperparameters ={∖σ}. The peak evidence may now be found more easily by numerically maximizing in (2.16) with respect to the remaining parameters . If a gradient-based algorithm is used, it is advantageous to have an analogous expression to (2.7) to give inexpensive derivatives. This can be found by differentiating (2.16) with respect to , making use of the results in (2.8), These are not the same as the derivatives in (2.7). As well as maximizing, we can also consider marginalizing over σ [16]. As we are marginalizing over a scale parameter we use the (improper) Jeffreys prior P(σ)=c/σ [17]. The result is equal to the maximized form, up to a multiplicative constant, As before, once the peak hyperlikelihood has been found, the Hessian at the peak position can aid in model comparison. In this case, the Hessian should be calculated using the second derivatives of . However, we may instead differentiate , as this differs only by a constant which will cancel when using the Hessian to compare two models. Differentiating (2.17) with respect to ′,[4] Again, these are not the same as the derivatives in (2.9). These expressions for the gradient and the Hessian of the hyperlikelihood, maximized or marginalized over , share the same advantages as the analogous expressions in §2: they may be evaluated in time once the hyperlikelihood itself has been evaluated in time.

Numerical results

In order to perform model comparison calculations between competing covariance functions, we must first specify at least two different covariance functions. We choose the two functions in (3.1) and (3.2), where (t,t′)≡(x,x′). These functions are both based on the periodic covariance function proposed by [2]. The first function k1 is the product of a single periodic component with time scale T1 and a simple compact-support polynomial covariance function [18] to describe any non-periodic component of the data. The choice of a compact-support covariance function is especially useful when working with large datasets; this is precisely the situation where the techniques described above are also designed to be of maximum benefit. The second function k2 includes an additional periodic component with time scale T2. In order to avoid double-counting in k2, we impose the constraint T2≥T1. Both covariance functions also include an uncorrelated noise term; we define this in such a way that σ remains an overall scale hyperparameter which can be maximized or marginalized over analytically as described in §3.2. The covariance functions are completely specified by the hyperparameters σ (overall scale), T (j=0,1,2; time scales) and l (j=1,2; smoothing parameters for the periodic components). The noise parameter σ could also be taken to be a hyperparameter; instead, for simplicity, we here take σ to be fixed. As σ appears in k multiplied by the overall scale, σ, fixing σ is roughly equivalent to specifying a fixed fractional error. We want to perform model comparison using the Laplace approximation outlined previously. This technique requires reparametrizing the covariance function such that the hyperpriors are flat. For the time-scale hyperparameters, which are dimensionful, we choose to use the scale-invariant Jeffreys prior, P(T)∝1/T. This prior is improper if the range of T is , so we restrict the range to (δt,ΔT), where δt and ΔT are respectively the smallest and largest separations between the sampling points. If there was a time scale in the problem outside of this range, we would be unable to resolve it from the data. We now seek a transformation ϕ≡ϕ(T) to a new hyperparameter ϕ such that the prior is flat in this parameter, P(ϕ)=const. The conservation of probability gives a differential equation relating the two where the As are constants which we can set equal to 1. The range of these new hyperparameters is and . For the smoothness parameters l, we choose to use lognormal priors, , with mean μ=1 and variance . As before, we seek a transformation to some new hyperparameters ξ in which the prior is flat. The desired transformation is given by where ξ∈(−0.5,0.5).

Synthetic data

Shown in figure 1 are realizations of GPs with covariance functions k1 and k2.[5] In order to perform test model comparison calculations, a realization of the k2 GP with n points was drawn and analysed using both the k1 and k2 covariance functions. For each covariance, the peak hyperlikelihood was found by numerically maximizing (2.14) using a conjugate gradient method, making use of the gradient in (2.17). The hyperevidence was estimated using (2.13) and the expression for the Hessian in (2.19); the results are summarized in table 1. To verify the accuracy of this estimate, the hyperevidence was also integrated numerically using MultiNest [19-21], which implements a nested sampling algorithm [22]. This was repeated for three different values of n (in the case n=100, the synthetic data are plotted in the right-hand panel of figure 1), and the results are also summarized in table 1.
Figure 1.

Realizations of the GPs k1(t,t′) and k2(t,t′) from (3.1) and (3.2) for values of t=1,2,3…,100 are shown in the (a,b) panels, respectively. The horizontal black lines indicate the length scales associated with the different terms in the covariance functions. The hyperparameters for k1 were chosen to be σ=1, ϕ0=3.5, ϕ1=1.5 and x1=0. The hyperparameters for k2 were chosen to be the same as for k1 and ϕ2=3 and x2=0. In both cases, the noise was fixed to σ=10−2.

Table 1.

A summary of the results of the analysis of synthetic data for three different-sized datasets. The first set of two columns is for a dataset drawn from the k2 covariance function and analysed with the k1 covariance function. The first column is the estimated hyperevidence using the Laplace approximation where is as given in equation (2.13), while the second is the numerically calculated hyperevidence. The second set of two columns shows results for the same data, but analysed with the k2 covariance function. The final pair of columns shows the log Bayes factor, , calculated using the approximate and numerical values for the hyperevidence.

nlnZestk1lnZnumk1lnZestk2lnZnumk2lnBestlnBnum
30−17.77−17.87±0.0818.82−17.73±0.09−1.050.14±0.12
100−20.17−20.17±0.10−19.22−19.22±0.110.950.95±0.15
300−49.94−50.12±0.11−40.21−40.36±0.139.739.76±0.17
Realizations of the GPs k1(t,t′) and k2(t,t′) from (3.1) and (3.2) for values of t=1,2,3…,100 are shown in the (a,b) panels, respectively. The horizontal black lines indicate the length scales associated with the different terms in the covariance functions. The hyperparameters for k1 were chosen to be σ=1, ϕ0=3.5, ϕ1=1.5 and x1=0. The hyperparameters for k2 were chosen to be the same as for k1 and ϕ2=3 and x2=0. In both cases, the noise was fixed to σ=10−2. A summary of the results of the analysis of synthetic data for three different-sized datasets. The first set of two columns is for a dataset drawn from the k2 covariance function and analysed with the k1 covariance function. The first column is the estimated hyperevidence using the Laplace approximation where is as given in equation (2.13), while the second is the numerically calculated hyperevidence. The second set of two columns shows results for the same data, but analysed with the k2 covariance function. The final pair of columns shows the log Bayes factor, , calculated using the approximate and numerical values for the hyperevidence. From table 1 it can be seen that as n is increased, the Bayes factors increasingly favour the more complicated covariance function (and in this case the correct covariance function from which the data was drawn). In almost all cases, the Laplace approximation gives a value which is in agreement at better than 2σ with the numerically integrated value . There is one exception which is highlighted in italic; this occurs for the most complicated covariance function (with the largest number of hyperparameters) and when the number of data points is smallest. In this situation, it would be expected that the posterior distribution on the hyperparameters may be highly multimodal and/or exhibit strong degeneracies (both of these expectations were confirmed by examining the posterior distribution on the hyperparameters returned by MultiNest). This exceptional case serves to highlight situations in which the Laplace approximation should not be trusted. The MultiNest posteriors in all other cases were verified to be well approximated by a single Gaussian mode. Figure 2 shows the posterior distribution for the parameters of k2 obtained from the largest (n=300) synthetic dataset.
Figure 2.

The one- and two-dimensional marginalized posterior distributions on the hyperparameters of the k2 covariance function obtained from the largest (n=300) synthetic dataset. The posterior is well approximated as a normal distribution. Shown in the black curves in the one-dimensional marginalized posterior distributions along the diagonal are the normal approximations obtained by using the techniques described in §2 to maximize and find the Hessian. Using the Hessian to approximate the integral of this distribution (the hyperevidence) leads to an error of approximately 10% (table 1).

The one- and two-dimensional marginalized posterior distributions on the hyperparameters of the k2 covariance function obtained from the largest (n=300) synthetic dataset. The posterior is well approximated as a normal distribution. Shown in the black curves in the one-dimensional marginalized posterior distributions along the diagonal are the normal approximations obtained by using the techniques described in §2 to maximize and find the Hessian. Using the Hessian to approximate the integral of this distribution (the hyperevidence) leads to an error of approximately 10% (table 1). Our method of model comparison is proposed as a faster alternative to model comparison using numerically evaluated Bayes factors. Simply comparing the peak hyperlikelihood (marginal likelihood) values would also give a measure of the goodness of fit, but this tends to favour more complex models and incurs the risk of overfitting. More sophisticated methods of model selection exist in the literature (see [23,24] and references within), e.g. the comparison of models based on estimated predictive criteria [25-27], or the construction of a larger reference model and the subsequent selection of a simpler submodel with similar predictions [28,29]. A detailed numerical comparison to these methods is left for future investigation. The values in table 1, evaluated using MultiNest, required between 20 000 and 50 000 likelihood evaluations. The maximization routines typically took fewer than 100 likelihood evaluations to find the peak, and then one additional evaluation to calculate the Hessian and hence . In order to guard against the possibility of the maximization routines becoming trapped in local maxima, as opposed to the global maximum, the algorithm was run multiple times from randomly selected starting positions. The typical number of runs required to find the global maximum was approximately 10. After these duplicate runs are accounted for, the speed-up factor in calculating compared with was between 20 and 50 in all cases.

Tidal data from Woods Hole

In order to illustrate the effectiveness of the techniques described above on real data, we consider several tidal datasets of different sizes from Woods Hole, MA, USA [30].[6] We consider the mean sea-level offset recorded between 3 January and 15 June 2014, six lunar months, sampled at 2 h intervals (giving n=1968 data points). This is plotted in figure 3. We also consider a smaller subset of the data (the first lunar month), with n=328 data points.[7]
Figure 3.

Shown in the main figure are six lunar months of tidal height data (black), from which the lunar tidal cycle can be discerned. Shown in the inset plot are several days of the tidal data (black points), from which the daily cycles can be clearly seen. Overlaid in the inset plot are both GP interpolants (blue), which are identical on this time scale.

Shown in the main figure are six lunar months of tidal height data (black), from which the lunar tidal cycle can be discerned. Shown in the inset plot are several days of the tidal data (black points), from which the daily cycles can be clearly seen. Overlaid in the inset plot are both GP interpolants (blue), which are identical on this time scale. We interpolate the data using the two covariance functions in (3.1) and (3.2); these functions are well suited to the data, as we expect the sea level to contain harmonics of the various time scales associated with the daily, monthly and yearly cycles of the tides. For simplicity, we fix σ=10−2, which is the typical fractional error in the sea-level measurements. As in §3.1, we reparametrize the covariance functions so that the T have Jeffreys priors, and the smoothness parameters have lognormal priors. We use a conjugate gradient maximization algorithm with (2.17) and the Hessian in (2.19) to evaluate the volume in (2.13) and perform model comparison between the two covariance functions. For the smaller dataset, we find the time scale T1=(12.8±0.2) h with k1, which corresponds to the two main tides per day. With k2, we find the time scales T1=(12.44±0.07) h and T2=(24.3±1.0) h; the second time scale corresponds to the height difference between the first and second tides of the day. The two time-scale model is highly favoured with a log Bayes factor of 57.8. For the larger dataset, we find T1=(12.80±0.11) h with k1, and T1=(12.40±0.03) h and T2=(23.3±0.3) h with k2. In all cases, the (squared) errors are estimated using the diagonal components of the inverse Hessian; it can be seen that the time scales are more precisely measured for the larger dataset, as expected. The two time-scale model is even more conclusively favoured for the larger dataset, with a log Bayes factor of 538. We also find a number of subsidiary hyperlikelihood peaks associated with other time scales in the data, but all subsidiary peaks are strongly suppressed relative to the global peak (by at least of approx. 100) and so we expect our Bayes factor estimates to be robust. Sea-level data are known to contain a large number of different frequencies, which necessitates the use of harmonic analysis in tidal modelling; the number of constituents included in tide prediction calculations has increased from tens [31] to thousands [32] over the past century. Clearly any k2-like covariance function with fewer than 10 time scales is simplistic, but the construction of a more detailed tidal model is beyond the scope of this paper.[8] The number of evaluations of (2.13) needed to obtain these results was comparable with the numbers for the synthetic data discussed in §3.1. However, each evaluation here was more expensive (approx. 10 s) due to the size of the dataset. Based on the speed-ups found in §3.1, it would be expected that multinest would take up to approximately one week to calculate the Bayes factor. Shown in the inset plot in figure 3 are the two interpolants from k1 and k2 for the larger dataset, which both perform equally well on the time scale of one week. These interpolants, which are the result of the regression analysis, may be used to estimate the tidal height at a time where a measurement is not available.

Summary

We have described some simple ways in which the computationally expensive training stage of implementing GPR can be accelerated. The analytic maximization of the hyperlikelihood over a single-scale hyperparameter of the covariance function aids in speeding up the maximization of the hyperlikelihood by reducing the dimensionality of the problem; the advantages of this will be most keenly felt in (common) problems where relatively simple covariance functions are used. Meanwhile, the analytic evaluation of the Hessian matrix, either in the manner of (2.9) or (2.19), aids in speeding up the process of model comparison between different types of covariance function. We have successfully demonstrated these techniques on a synthetic dataset where the data was drawn from one of two covariance functions under consideration. In the case of the synthetic data, the size of the dataset was limited to fewer than 300 points, so that the results could be verified by using the MultiNest algorithm to numerically sample and integrate the posteriors. We also demonstrated the techniques by applying them to a larger real dataset of mean sea-level measurements, where the full MultiNest calculation would have taken too long to perform. It is to be hoped that these techniques will aid in the wider application of GP methods to larger datasets.
  3 in total

1.  An invariant form for the prior probability in estimation problems.

Authors:  H JEFFREYS
Journal:  Proc R Soc Lond A Math Phys Sci       Date:  1946

2.  Novel method for incorporating model uncertainties into gravitational wave parameter estimates.

Authors:  Christopher J Moore; Jonathan R Gair
Journal:  Phys Rev Lett       Date:  2014-12-15       Impact factor: 9.161

3.  Efficient Gaussian process regression for large datasets.

Authors:  Anjishnu Banerjee; David B Dunson; Surya T Tokdar
Journal:  Biometrika       Date:  2013-03       Impact factor: 2.445

  3 in total
  6 in total

1.  Scalable Nonparametric Prescreening Method for Searching Higher-Order Genetic Interactions Underlying Quantitative Traits.

Authors:  Juho A J Kontio; Mikko J Sillanpää
Journal:  Genetics       Date:  2019-10-04       Impact factor: 4.562

2.  Quantifying vegetation biophysical variables from the Sentinel-3/FLEX tandem mission: Evaluation of the synergy of OLCI and FLORIS data sources.

Authors:  Charlotte De Grave; Jochem Verrelst; Pablo Morcillo-Pallarés; Luca Pipia; Juan Pablo Rivera-Caicedo; Eatidal Amin; Santiago Belda; José Moreno
Journal:  Remote Sens Environ       Date:  2020-12-15       Impact factor: 13.850

3.  A Survey of Active Learning for Quantifying Vegetation Traits from Terrestrial Earth Observation Data.

Authors:  Katja Berger; Juan Pablo Rivera Caicedo; Luca Martino; Matthias Wocher; Tobias Hank; Jochem Verrelst
Journal:  Remote Sens (Basel)       Date:  2021-01-15       Impact factor: 5.349

4.  Monitoring Cropland Phenology on Google Earth Engine Using Gaussian Process Regression.

Authors:  Matías Salinero-Delgado; José Estévez; Luca Pipia; Santiago Belda; Katja Berger; Vanessa Paredes Gómez; Jochem Verrelst
Journal:  Remote Sens (Basel)       Date:  2021-12-29       Impact factor: 5.349

5.  Optimizing Gaussian Process Regression for Image Time Series Gap-Filling and Crop Monitoring.

Authors:  Santiago Belda; Luca Pipia; Pablo Morcillo-Pallarés; Jochem Verrelst
Journal:  Agronomy (Basel)       Date:  2020-04-27       Impact factor: 3.949

6.  Green LAI Mapping and Cloud Gap-Filling Using Gaussian Process Regression in Google Earth Engine.

Authors:  Luca Pipia; Eatidal Amin; Santiago Belda; Matías Salinero-Delgado; Jochem Verrelst
Journal:  Remote Sens (Basel)       Date:  2021-01-24       Impact factor: 5.349

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.