Venkat Roy1, Andrea Simonetto2, Geert Leus3. 1. NXP Semiconductors, High Tech Campus 46, 5656 AE Eindhoven, The Netherlands. venkat.roy@nxp.com. 2. Optimisation and Control group, IBM Research Ireland, Dublin 15, Ireland. andrea.simonetto@ibm.com. 3. Faculty of Electrical Engineering, Mathematics and Computer Science, Delft University of Technology, Mekelweg 4, 2628 CD Delft, The Netherlands. g.j.t.leus@tudelft.nl.
Abstract
We propose a sensor placement method for spatio-temporal field estimation based on a kriged Kalman filter (KKF) using a network of static or mobile sensors. The developed framework dynamically designs the optimal constellation to place the sensors. We combine the estimation error (for the stationary as well as non-stationary component of the field) minimization problem with a sparsity-enforcing penalty to design the optimal sensor constellation in an economic manner. The developed sensor placement method can be directly used for a general class of covariance matrices (ill-conditioned or well-conditioned) modelling the spatial variability of the stationary component of the field, which acts as a correlated observation noise, while estimating the non-stationary component of the field. Finally, a KKF estimator is used to estimate the field using the measurements from the selected sensing locations. Numerical results are provided to exhibit the feasibility of the proposed dynamic sensor placement followed by the KKF estimation method.
We propose a sensor placement method for spatio-temporal field estimation based on a kriged Kalman filter (KKF) using a network of static or mobile sensors. The developed framework dynamically designs the optimal constellation to place the sensors. We combine the estimation error (for the stationary as well as non-stationary component of the field) minimization problem with a sparsity-enforcing penalty to design the optimal sensor constellation in an economic manner. The developed sensor placement method can be directly used for a general class of covariance matrices (ill-conditioned or well-conditioned) modelling the spatial variability of the stationary component of the field, which acts as a correlated observation noise, while estimating the non-stationary component of the field. Finally, a KKF estimator is used to estimate the field using the measurements from the selected sensing locations. Numerical results are provided to exhibit the feasibility of the proposed dynamic sensor placement followed by the KKF estimation method.
Tracking the spatio-temporal evolution of any field using a limited number of homogeneous/heterogeneous sensors with a desired accuracy is one of the most common applications of wireless sensor networks (WSNs) [1,2]. Different types of environmental, geophysical and biological processes exhibit complicated spatial as well as temporal variability. Spatial and temporal variability of a spatio-temporally stationary physical field can be modelled by its correlation over space and time [3]. If the field is non-stationary then a suitable dynamic model can be used to model the spatio-temporal evolution of the field [3]. If the field exhibits both a stationary and non-stationary behavior over space and time then the field can be dynamically monitored by the combination of kriging [3] and Kalman filtering, i.e., a kriged Kalman filter (KKF) [4] or space-time Kalman filter [5]. The key idea behind the KKF is the liaison of kriging [3] and Kalman filtering. The unknown physical field is modelled as a combination of a non-stationary (capturing the dynamics) and a stationary (capturing the low magnitude spatial effects) stochastic component. Assuming that the dynamics of the non-stationary component and the second-order statistics of the stationary component (e.g., covariance structure) are perfectly known, KKF jointly estimates both of these field components using the spatial observations at every time instant. The KKF paradigm has been used for different applications like wireless communications (e.g., spectrum sensing [6] and path delay estimation [7]) and field estimation [5]. From a practical perspective, KKF can also be used for air pollution level forecasting applications [8]. Also, a distributed implementation of KKF can be used for environment monitoring using a robotic sensor network [9].One of the important overheads of dynamic field estimation using a WSN is the lack of sufficient measurements at every time instant. This is related to the shortage of sensor life time, availability of bandwidth, and other resource-related economical constraints. In such scenarios, we need to efficiently place/move the available sensors into the most informative locations over space and time. One notable approach for sensor placement in Gaussian processes is based on exploiting the submodularity of the mutual information between the sensing locations [10]. Submodularity of the frame potential and related greedy algorithms for sensor placement are proposed in [11]. Dynamic sensor scheduling is a well-cultivated topic in the fields of signal processing as well as control theory [12,13]. Some recent notable contributions are [14,15,16]. Prior knowledge regarding the correlation of the field over space and time can be exploited in a multi-layer design of sensor networks [17]. Selecting the most informative sensing locations can be treated as a sensor selection problem, which can be formulated as a convex optimization problem [18]. This can be solved for linear [19] as well as non-linear measurement models [20]. Sparsity-promoting approaches for sensor placement are also exhibited in [15,21], where the placement algorithm is formulated using the alternating direction method of multipliers (ADMM). In [22], a generalized sparsity-enforcing and performance-constrained sensor placement method is proposed, where the field can be either stationary or non-stationary. The aforementioned method can be implemented for a single snapshot or multiple snapshot ahead sensor placement and for a general class of spatio-temporal covariance matrices, which can either be ill-conditioned or well-conditioned. Seminal contributions on the convex formalism of sensor selection (like [18]) assume that the measurement noise components are spatio-temporally uncorrelated. However, this can be an unrealistic assumption in some practical scenarios [23]. However, even in those scenarios, it has been shown that the sensor selection problem can be formulated as a convex optimization problem [16,24].In this work, we develop a unified framework of sensor placement followed by a KKF estimator to dynamically monitor a physical field that exhibits both stationarity and non-stationarity over space and time. In the first step, we select the most informative locations to deploy/move the sensors and in the second step we estimate the field by using the measurements from those selected locations. The key contributions can be summarized as follows,The performance metrics to estimate the stationary as well as the non-stationary components of the field are represented in closed form as an explicit function of the sensor location selection vector.The aforementioned analytical formalism tackles two important issues in the sensor placement step. First, the developed method takes care of the fact that the estimation of the non-stationary component of the field involves the stationary component of the field as a spatially correlated observation noise. Second, the proposed method is applicable for a general class of spatial covariance matrices of the stationary component of the field, even when they are ill-conditioned or close to singular [25].The proposed sensor placement problem is formulated in a way that minimizes a cost function that involves the sum of the mean square error (MSE) of the stationary and the non-stationary component of the field as well as a spatial sparsity enforcing penalty. The overall optimization problem also satisfies a flexible resource constraint at every time instant.One of the aspects that distinguishes the proposed sensor placement method from the prior works in sensor placement for environmental field estimation is the specific statistical nature of the unknown physical field, which yields an additive coupling of stationary and the non-stationary components. Secondly, we develop a unified framework for the efficient utilization of the spatio-temporal variability of the field in order to design an opportunistic sensor placement method using a convex approach. We develop a parsimonious sensor placement algorithm followed by a KKF estimator, which can be used to dynamically monitor a general class of environmental fields (based on the assumed process model and spatial statistics of the field components). However, the developed approach is similar to [22] in terms of the primary measurement model, which is considered to be linear and underdetermined. We emphasize that the proposed technique is a model-based centralized sensor placement method, where we resort to the Bayesian estimation philosophy. We assume that the available prior statistical knowledge regarding the unknown physical field like spatial correlation information and dynamics are perfectly known a priori. It is also assumed that for the current centralized setup the communication range of the sensors are sufficient to communicate with the fusion center, which can be located inside/outside the given service area.Notations: Matrices are in upper case bold while column vectors are in lower case bold. The notation is the ()-th entry of the matrix , is the i-th entry of the vector , and denotes the trace of , i.e., the sum of the diagonal elements of . The notation is defined as the set of indices of the non-zero entries of , while and are the diagonal matrix with diagonal and the main diagonal of the matrix , respectively. An identity matrix of size is denoted by . The notation is the transpose operator, is the estimate of , and is the norm of . Vectors of all zeros and ones of length N are denoted by and , respectively. An all zero matrix of size is given by . The set of symmetric matrices of size and the set of symmetric positive-definite matrices of size are denoted by and , respectively.
2. Signal Modelling and Problem Formulation
2.1. Measurement Model
Let us denote the spatially continuous field by , at any discrete time index t and location . We assume that the entire service area of interest is uniformly discretized over N pixels, where we would like to estimate the field intensities. The field intensities of the N pixels at time t are represented by . It is assumed that the field intensity is the same everywhere within a pixel, and it can be represented by , where is the centroid of the j-th pixel, with . We consider a linear underdetermined measurement model
where is the non-stationary component of the field and is a stationary component of the field capturing the non-dynamic spatial effects. It is assumed that and are mutually uncorrelated.At any time t, the measurements are given by collected from () sensing locations (pixels) of the entire service area. The time-varying sensing or measurement matrix selects measurements from N field locations. The measurement matrix is constructed by removing the zero rows of , where is a sensor location selection vector. If , then the j-th pixel is selected (not selected) for sensor placement. Based on this, at any time t, the number and the constellation of the selected sensing locations are given by and , respectively. Using the considered construction of , we have the relationsNote that the type of measurement matrix used in this work is similar to an incidence matrix, which can be viewed as a flexible data collection method using heterogeneous sensing equipments. In practice, when different types of sensing modalities are used, we may not know the process by which any of the sensors gathers its measurement but only its recorded value is important. Also, we rigorously exploit the property of the structure of mentioned in (3), later in this paper.The error incurred by the measurement process is modelled through , which is uncorrelated with both and , respectively. The spatio-temporally white measurement noise is characterized by .
2.2. Modelling of the Spatial Variability
The spatial effects of the field are modelled through a spatially colored yet temporally white discrete random process , where is the mean and is the spatial covariance matrix of . We assume that the process is spatially second-order stationary as well as isotropic, which means that
where [3]. Note that here we follow the same spatial discretization as mentioned in Section 2.1. There are several empirical as well as parametric model-based approaches to model the spatial covariance. In this work, we assume that the spatial covariance function is given by a simple squared exponential function:
where is the parameter controlling the strength of the spatial correlation. The covariance function mentioned in (6) is plotted in Figure 1a for increasing values of the pairwise distance between the centroids of the pixels, i.e., and the parameter . Note that the aforementioned covariance function belongs to the family of Matérn covariance functions [3], which are widely used to model the spatial variability of a field in geostatistics and environmental sciences.
Figure 1
(a) Squared exponential covariance function for different values of (variance ); (b) Service area with candidate sensing locations; (c) Spatial covariance matrix.
Using the squared exponential covariance function, the elements of the spatial covariance matrix () can be constructed by the Relation (5). Let us consider a service area with pixels. The centroids of these 36 pixels, which are also the candidate locations for sensor deployment are shown in Figure 1b. These centroids are indexed from the top left to the bottom right. The elements of are shown in Figure 1c. Note that based on the nature of the covariance Function (6), the spatial covariance matrix is symmetric and based on the constellation of the candidate sensing locations (Figure 1b), is also a block Toeplitz matrix. We assume that and are perfectly known a priori.
2.3. State Model
The spatio-temporal evolution of the non-stationary component of the field, i.e., , can be modelled by the following state modelHere, the time-varying state transition/propagator matrix is given by . The process noise vector is assumed to be characterized by . The elements of the state transition matrix act as spatial redistribution weights for for the temporal transition from to t [3]. Note that this spatial redistribution can be dependent on the temporal sampling interval. We model the elements of by using a parameterized Gaussian kernel function
where and the spatio-temporally varying translation and dilation parameters are represented by , and , respectively. The scalar is a scaling parameter to avoid an explosive growth of , i.e., to keep the maximum eigenvalue of below 1. The simplest form of the state model (7) is defined by , which is similar to a Gaussian random walk model. This corresponds to the Gaussian kernel Function (8), with and , and with . In this work, we assume that the state transition matrix , whose elements are parameterized by and through the Function (8) is perfectly known a priori.From a practical point of view, the aforementioned Gaussian kernel-based spatio-temporal evolution can be used for rainfall prediction [26]. This modelling approach incorporates some physical properties of environmental fields, such as diffusion, advection etc. [26], so it can also be used for modelling the propagation of a general class of environmental fields (e.g., pollutants, aerosol movements).
2.4. Main Problem Statement
The main problem is to design an optimal sensor placement pattern, i.e., to design , whose support gives the optimal locations to deploy the sensors. At any t, the design goal is to minimize the estimation error for both the stationary and the non-stationary components of the field as well as to enforce sparsity in , i.e., to reduce the number of required sensing locations. If the estimation error of the stationary and non-stationary components of the field can be represented by a single performance metric , the sensor placement problem can be represented byAt any t, and denote the lower bound on the number of available sensors, and a given budget on the maximum number of available sensors, respectively. Sparsity is enforced through a sparsity-promoting penalty, i.e., an norm of in the second summand of the cost Function (9a) with a time-varying regularization parameter controlling the sparsity of the elements of . A detailed description regarding the structure of the objective function and the importance of the constraints in the optimization problem (9) are discussed in Section 3.1.
2.5. Simple KKF Estimator and Estimation Error Covariance
Using the measurement and state models of (1) and (7), respectively, the state estimate , for , can be computed following the lines of a standard KKF [5,6]. First, a simple Kalman filter is used to track the dynamic component , where the stationary component is interpreted as a noise term. In this case, the measurement model is given by
where , , and . Furthermore, , and , with . It can be easily shown that and are mutually uncorrelated as it is already assumed in Section 2.1 that is mutually uncorrelated with and , respectively. Now, using the state model of (7) and the measurement model of (10), the non-stationary component can be estimated following the lines of a simple Kalman filter [27]. In this case, the recursive state estimate at time t is given by
where the Kalman gain can be expressed asThe MSE matrix of the estimate at time t is given by , which is related to the MSE matrix of the estimate at time , i.e., , by the recursive relation [27]In the next stage, the estimate of in (11) is used to compute the stationary component using kriging. In spatial statistics, the intensity of an environmental field in an unknown location can be interpolated using a variogram model [3,4]. For a spatially stationary field, this variogram can be expressed as a covariance [3]. In this way kriging can be viewed as a simple linear minimum mean square error (LMMSE) interpolator [27]. The linear model is given by and the related estimator has the form
where we use the prior information . The MSE matrix [27] of the estimate , i.e., is given byFinally, the overall field estimate at time t is given by .
2.6. Performance Metrics as a Function of
In this section, we express the MSE matrices, i.e., and as functions of . First of all, we mention some facts regarding the structure of the error covariance matrices presented in the Expressions (13) and (15).It should be noted that the measurement noise in (10) is correlated over space through the off-diagonal elements of . Due to this fact, sensor selection approaches using the standard convex framework like [18,20,22], i.e., designing a by directly optimizing the Expression (13) is difficult due to the presence of the off-diagonal elements of . It should also be noted that the expression of is a function of the measurement matrix and thus the selection vector . However, we do not encounter this issue in the performance metric to estimate the stationary component , i.e., (15), as the measurement noise is assumed to be spatially uncorrelated in this case.In the expression of , i.e., (15), we assume that is well-conditioned, i.e., accurately invertible. However, this may not be the case in some scenarios. The condition number of strongly depends on the correlation of the field, spatial sampling distance, grid size etc. [25]. The variation of the condition number of with different values of for both and is plotted in Figure 2a. It is seen that for a higher resolution or a strong spatial correlation, the spatial covariance matrix becomes increasingly ill-conditioned and thus close to singular. In such circumstances, we cannot compute the estimation error covariance matrix using the Expression (15). In that case, can be computed using the alternate expression of (15) given by
which is obtained using the matrix inversion lemma (MIL). It should be noted that the alternative expression of the MSE can be used to compute the MSE (without inverting ), but it is difficult to express it as an explicit function of .
Figure 2
(a) Plot of the condition number of vs. with different number of candidate sensing locations (N); (b) MSE of the estimate of vs. for different numbers of candidate sensing locations (N); ; ; The spatial resolution is increased by representing one pixel of Figure 1a by 4 pixels.
In Figure 2b, we plot for the best case, i.e., the MSE with all the pixel centroids equipped with sensors ( or ) for different values of , and for two different spatial resolutions ( and ) of the same square km service area. It is seen that decreases as the strength of the correlation is increased by increasing .To circumvent the effect of ill-conditioning as well as to handle the correlated measurement noise in the expression of , we propose the following approach. We start by introducing the substitution
where is a well-conditioned matrix and . More specifically, the range of is considered to be . Substituting , we can represent the measurement error covariance matrix of (10) as, , where and where we used . Substituting in (13), the MSE matrix for the estimate of the non-stationary component is given byUsing the MIL, we have the following matrix identity
from which we can derive
where we used . Substituting (20) in (18) we obtain the following expression for .Next, substituting in the inverse of the right most term of (16) and using , we obtainSubstituting the identity (20) into (22), we obtain the following expression of .Note that, the expression of (23) avoids the inversion of an ill-conditioned . Here, we only need to invert the well-conditioned .In this work, we consider the overall MSE as a performance metric for sensor placement, i.e., as mentioned in (9a). This is given by
where , , , and . Note that the matrices , , , and are all independent of . To model and as positive definite matrices we need .The performance metric derived in (24) incorporates the MSE matrices of the estimates of the non-stationary () as well as the stationary () component of the field, as explicit functions of the sensor location selection vector . Note that a formulation similar to (23), for the computation of the MSE matrix as a function of is proposed in [22], where the field is considered to be either purely stationary or non-stationary.
3. KKF with Sensor Placement
In this section, we relax and reformulate the proposed sensor placement problem (9) as a semidefinite programming (SDP) problem. Then we present the overall KKF estimator followed by the sensor placement to dynamically monitor the field using only the measurements from the selected sensing locations.
3.1. Sensor Placement Problem as an SDP
Solving for the best subset of sensing locations is a combinatorially complex problem. However, it can be relaxed to a convex problem [18,19,20]. As discussed in Section 2.1, the sensor location selection vector acts as a weighting vector for all the N candidate pixels. Following the main optimization problem, i.e., (9), a sparsity-enforcing, low estimation error, and resource-constrained design of can be obtained by solving
where the expression of is given by (24). Here, we have relaxed the non-convex Boolean constraint of (9) to a convex box constraint . The resource constraint of (25b) is affine and thus convex. Some comments regarding the formulation of the proposed sensor placement problem of (25) are presented next.First of all, let us consider the non-convex version of the optimization problem of (25) with . This is given asIn this case, the MSE cost will be minimum, i.e., the best estimation performance is achieved, when we select the maximum number of available candidate locations or in other words, when . Then, there is no way to reduce the number of selected locations below and the constraint becomes redundant. In the aforementioned case, it is difficult to reduce the number of selected sensing locations below .Notice that, dropping the resource constraint (25b) and increasing will reduce the number of selected sensing locations. However, there is no explicit relation between and , i.e., it is difficult to directly control the resource allocation (i.e., ) through .We mention that the proposed formulation of (25) is not a direct MSE minimization problem but it attains a specific MSE along with enforcing sparsity in spatial sensor location selection through the second summand of (25a). The sparsity enforcement is lower bounded by the minimum number of sensing locations to be selected at any t, i.e., . It should be noted that for an arbitrary selection of , the minimum number of selected sensing locations will always be .Lastly, it should be noted that a sparsity-enforcing design of can be achieved by retaining only the second summand of the objective function of (25a) and using a separate performance constraint given as [20,22]. The desired performance threshold can be time-varying or independent of t based on the application. However, in many practical scenarios, it could be difficult to set the performance threshold a priori for every t.Based on the aforementioned arguments, we advocate the proposed design approach (25) that lowers the MSE along with enforcing sparsity in sensor placement satisfying a flexible resource allocation constraint.After solving (25), we obtain which can be converted to a Boolean selection vector . This can be performed by either deterministic or stochastic rounding procedures as discussed below.The simplest approach could be to set the non-zero entries of to 1. However, there can be a huge difference between the magnitudes of any two non-zero elements in . Considering the fact that the indices of the high magnitude (close to 1) elements of signify a more informative sensing location, can be sorted in ascending order of magnitude [18] and a selection threshold () can be selected based on the magnitudes of the elements of the sorted . The entries of the Boolean selection vector can be computed as if else , for .Another approach could be a stochastic approach, where every entry of is assumed to be the probability that this sensing location is selected at time t. Based on this, multiple random realizations of are generated, where the probability that is given by , for . Then the realization that satisfies the constraints and minimizes the estimation error, i.e., is selected [20].Let us now transform the optimization problem of (25) into an SDP. From the expression of (24), it is clear that minimizing w.r.t. is equivalent to minimizing the expression as the matrix is independent of . In the first step, we represent the optimization problem of (25) in an epigraph form ([28], p. 134), ([18], Equations (25) and (26)) which is given by
where we introduce the auxiliary variables and . We notice that the epigraph form (27) is well-posed since by choosing in (17) the matrix is always positive definite and symmetric. In addition, also the matrix is also positive definite by construction as derived in (18)–(21).The epigraph form (27) is not a strictly convex program, in the sense that there are multiple and matrices that achieve the minimal cost value. This is due to the inequality constraints of (27b) and (27c). At optimality, the eigenvalues of and must be equivalent to their lower bounds in (27b) and (27c). Hence, an optimizer of the problem is and .We proceed by simplifying the constraint (27b). Let us introduce another auxiliary variable and substitute (27b) with two constraintsWith this in place, the optimization Problem (27) can be formulated asIt can be claimed that the optimization problem (30) is equivalent to (27) given that it yields a decision variable with the same optimal cost of (27) . To prove this, let us choose an arbitrary say . For a fixed yet arbitrary verifying (30e), the optimization problem (30) minimizes both and . This means that due to (30b) it minimizes also : in fact, as the lower bound for is minimal if the positive definite matrix is maximal, that is is minimal. Therefore, must attain its lower bound. As mentioned earlier, there are multiple optimizers, yet one is . In addition, at optimality, as well. The same reasoning holds also for , which at optimality is . Since this reasoning is valid for any feasible , it is also valid for an optimal one and therefore the equivalence claim follows. It should be noted that a similar argument was also presented in [24], where only the issue of correlated measurement noise is considered.Using the Schur complement lemma the constraints (30b) and (30c) can be equivalently represented by the linear matrix inequalities (LMI):The constraint (30c) can be equivalently represented by an LMI using the Schur complement [28]. In other words, using the fact that , we obtainFinally, an SDP representation of the overall optimization problem of (27) can be expressed asThe solution of the aforementioned SDP is a selection vector .
3.2. Spatial Sensor Placement for Stationary Field Estimation
Let us consider the effect of the stationary component of the field for any t. In this case, we consider that . In this case, the measurement model of (1) is given by . Exploiting the prior information regarding , i.e., an LMMSE estimator of can be presented by . The performance of the aforementioned estimator is given by the MSE matrix . Considering the fact that can be ill-conditioned, following the formulation of (23), the expression of can be expressed as a function of as
where , , and . Note that the matrices , , and are all independent of . Considering and following the same SDP formulation of Section 3.1, the proposed sensor placement problem of (9) can be represented asThe optimization problem of (35) gives the spatial sensor placement pattern for any snapshot t, when the field is stationary over space. However, if the field is also temporally stationary then the sensor placement problem of (35) can be extended to blocks of multiple snapshots. In this case, the performance metric can be computed using the same approach as [22]. In the simulation section, we show the effects of spatial correlation on sensor placement.
3.3. Sparsity-Enhancing Iterative Design
In order to eschew the effect of the magnitude dependencies of the elements of , we individually weigh each element of . In this case, we consider a vector form for the regularization parameter : . The weight associated to the each element of is the corresponding element of . We iteratively refine the weighting vector in the minimization term of the problem (33) [29]. Using this approach, higher weights are applied on the smaller elements of to push them towards 0 and the magnitudes of the larger elements are maintained by applying a smaller weight. In this way, a sparser solution can be obtained compared to the standard sparsity-promoting method. The iterative algorithm can be summarized asInitialize, weight vector , , and maximum number of iterations I.for, for everyend;set.After solving the above algorithm, we still obtain . We convert this to a Boolean selection vector using a deterministic/stochastic rounding method as mentioned in Section 3.1.
3.4. KKF Algorithm with Sensor Placement
The informative locations to deploy/move the sensors at any t is denoted by , where . The noisy measurements collected from the aforementioned locations are stored in . The sensing matrix is constructed by removing the all-zero rows of at every t. This measurement matrix is used for the estimation of the non-stationary and the stationary components by (11) and (14), respectively. Then the overall field estimate at time t is computed by . Note that the estimation steps, i.e., (11) and (14) do not require the computation of the inverse of . The error covariance of the non-stationary component can be updated by (13), which also does not require the inverse of . At every t, the overall estimation performance can be computed by the expression of (24). The best case performance, i.e., the performance with all the locations selected can also be computed by the expression of (24) by using .In many practical environmental fields (such as rainfall), the field is generally non-negative. To achieve a non-negative estimate at every t, the estimates of the stationary and non-stationary components can be projected onto the non-negative orthant, i.e., the negative values are set to zero. This is obtained by adoptingHowever, in this case, the performance metrics and are only the approximations. The overall sensor placement followed by a KKF algorithm is presented in Algorithm 1.Initialize: , , , .Given: , , , , I, , , .forPrediction using the state model.Sensor placement: iterative sparsity-enhancing design of (Section 3.3).Correction: estimation of and using the measurements from the selected sensing locations.Overall KKF estimate:
.Covariance update.end for
4. Simulation Results
In this section, we perform some numerical experiments to exhibit the practicality of the developed sparsity-enforcing sensor placement followed by the KKF estimation method. We select a service area of square km with 1 square km spatial resolution as illustrated in Figure 3. The spatial distribution of the non-stationary component at time , i.e., , is generated by the following exponential source-field model
where K is the number of field-generating points/sources. The parameters , , and are the location, amplitude, and the spatial decaying factor of the k-th source at time . Based on this function, we generate the non-stationary component of the field at time , i.e., using the parameters , , , and . The spatial distribution of in the specified service area is shown in Figure 3.
Figure 3
Field distribution at with a single source: , , , .
The state model of the non-stationary component is modelled by (7). The state transition matrix is modelled by the Gaussian kernel function given by (8). For the sake of simplicity, we consider a spatially invariant translation parameter and spatio-temporally invariant dilation parameters given as and , respectively, for . The elements of the state transition matrix are given byThe spatio-temporal evolution of the true value of the field, i.e., is generated in the following two ways.In the first case, we consider a pure advective process, i.e., we select a very low dilation parameter given by for all and . It is assumed that the temporal resolution is 1 min. The translation vectors, i.e., , are assumed to be changing every t as , , , , , , , and . It is assumed that at , is generated by the source as shown in Figure 3. The different states of for are generated by the state model of (7). The spatially colored yet temporally uncorrelated process noise is characterized by , where . The stationary component is modelled by . The parameters of the squared exponential covariance function of (6) are given by and . Note that increasing the value of , the field becomes spatially more correlated and the condition number of increases. However, as mentioned earlier, our proposed formulation, i.e., both the selection and the estimation, does not involve the inversion of . A highly spatially correlated is considered in the next case. For the first case, the true field for can be simulated as shown in Figure 4.
Figure 4
Spatio-temporal evolution of in a square km area; spatial resolution: square km; time varying for ; strength of spatial correlation: .
In the second case, we consider for all and the translation parameters are fixed as for , and no translation for the last 4 snapshots, i.e., for . The state of at is the same as before. The scaling parameter is given by . The process noise is the same as before. In this case, we assume that the stationary component is spatially more correlated than the last case. The parameters of the covariance Function (6) are taken as and , which generates an ill-conditioned (Figure 2a). Using these, the true field, i.e., for is shown in Figure 5.
Figure 5
Spatio-temporal evolution of in a square km area; spatial resolution: square km; time varying for ; strength of spatial correlation: .
4.1. Sensor Placement Followed by Field Estimation Using KKF
We select the optimal sensing locations and use them to estimate the field for snapshots for the two different scenarios of the spatio-temporal evolution of the field, as mentioned in the previous section. We use the same service area shown in Figure 1b, where the centroids of the pixels are the candidate sensing locations. We assume that the measurement noise variance is given by . We solve the optimization problem of (36) with the parameters and . The weighting vectors are initialized as . The resource constraints are given as and for all t. To extract the Boolean solution from , we adopt the randomized rounding method. We use the software CVX [30] (parser CVX, solver SeDuMi [31]) to solve the SDP problem (36).Following the above simulation setup, the selected sensing locations for the first and the second scenario are shown in Figure 6a,b respectively for the 8 snapshots. The indices of the pixel midpoints are the same as in Figure 1b (vertical axis). The main observations from the selected locations are listed below.
Figure 6
(a) Selected sensing locations to estimate the field with the first scenario of the true value; (b) Selected sensing locations to estimate the field with the second scenario of the true value.
First of all, it is clearly seen that the selected sensing locations depend on the dynamics. Note that Figure 6a gives the optimal placement pattern, when is changing every t (different on every t). Figure 6b shows the optimal sensing locations when we have the same for () and another for ().When is changing every t, i.e., the spatio-temporal evolution of the field is guided by the time-varying spatial translation parameter (Figure 1b), the optimal selection pattern also depends upon this translation (Figure 6a).In the second scenario, we have assumed a very low and fixed translation, i.e., for the first 4 snapshots and , i.e., no translation, for the last 4 snapshots (Figure 5). It is seen that almost the same set of sensors are selected in the last 4 snapshots of Figure 6b. In general, when is not changing with time, the estimation error of the non-stationary component reaches a steady state after a number of snapshots and the same set of sensors are selected every t.The overall estimation performance using the measurements from the selected locations of Figure 6a,b is shown in Figure 7a,b, respectively. In these figures, we exhibit the pixel-wise comparison of the estimates for snapshots, i.e., the estimation performance of pixels. We initialize the KKF iterations with and at .
Figure 7
(a) Comparison of the KKF estimate () with the true value () (Figure 4) with the measurements from the selected locations shown in Figure 6a; (b) Comparison of the KKF estimate () with the true value () (Figure 5) with the measurements from the selected locations shown in Figure 6b.
4.2. Performance Analysis
We compare the estimation performance of the developed sensor placement method by comparing the performance metric, i.e., with a random sensor placement (with the same , i.e., as for the developed method) and with the best case performance (i.e., or ). For the random placement, we generate 100 different realizations of at every t with the same as for the proposed method. Then is computed for every and their average is considered. Similarly, we compute the best case performance, i.e., for every t and in this case is updated with . We use the same set of parameters as mentioned in the first case of Section 4.1. Only the resource allocation constraint is simplified as , i.e., we fix that only 15 sensing locations will be selected every t. The performance comparison is shown in Figure 8. It is seen that the proposed approach slightly outperforms the random placement. However, the random placement of sensors does not optimize any performance criterion.
Figure 8
Comparison of the performance metric for the proposed method, random placement and the best case (only 15 sensors are selected on every t).
4.3. Spatial Sensor Placement for Stationary Field Estimation
In this section, we show the effects of different spatial correlation patterns on sensor placement assuming the field is purely stationary. We solve the optimization problem of (35), for two different spatial covariance matrices (). In the first case, we consider that is generated by the squared exponential Function (6) with and (Figure 9a). In the second case, we consider a randomly generated (Figure 9b). The resource allocation constraint is the same as before, i.e., , and . We solve the optimization problem of (35), with the iterative approach of (36) with the same parameters as mentioned in the previous section. The selected locations (marked by black squares where the blue dots are the candidate locations as shown in Figure 1b) to deploy sensors are shown in Figure 10a,b for the spatial covariance matrices shown in Figure 9a,b, respectively.
Figure 9
(a) Spatial covariance matrix generated by the squared exponential function (); (b) Randomly generated spatial covariance matrix.
Figure 10
(a) Sensor placement pattern for the as shown in Figure 9a; (b) Sensor placement pattern for the as shown in Figure 9b.
First of all, it is observed that the spatial distribution of the optimal sensing locations depends upon the correlation pattern of the field. It is seen that when is generated by a squared exponential covariance (stationary) function then the optimal sensor placement pattern is more or less symmetrically and uniformly distributed over the entire service area. However, for a random spatial covariance matrix the optimal sensing locations do not follow any specific pattern.
5. Conclusions and Future Work
In this work, we have developed a sparsity-enforcing sensor placement followed by a field estimation technique using a KKF. The proposed methodology selects the most informative sensing locations over space and time in a specified service area of interest. Along with minimizing the estimation error, the developed method also economizes the sensor placement (in terms of resources) at every temporal interval. The salient features of the proposed method include handling a general class of spatial covariance matrices and tackling correlated measurement noise. Numerical analysis shows the feasibility of the method. The effects of the dynamics and spatial correlation of the field in spatio-temporal sensor placement are discussed with numerical experiments.In this work, we have considered the fact that the prior knowledge regarding the spatial variability and the dynamics are perfectly known a priori. In that case, the performance of a clairvoyant Kalman setup with Gaussian measurement and process noise is optimal. However, in many practical scenarios, the aforementioned spatio-temporal prior information may not be accurate and we require more information regarding the unknown field in the estimation step. Future research is envisioned to incorporate the effects of model imperfections in the developed method. Another future research area could be using distributed algorithms to apply the developed method for large scale sensor network applications. It will also be interesting to tailor the recent progress in time-varying optimization [32] to solve the SDPs in a tracking fashion, rather than at optimality at each sampling time.