| Literature DB >> 33265201 |
Yosra Marnissi1, Emilie Chouzenoux2,3, Amel Benazza-Benyahia4, Jean-Christophe Pesquet3.
Abstract
In this paper, we are interested in Bayesian inverse problems where either the data fidelity term or the prior distribution is Gaussian or driven from a hierarchical Gaussian model. Generally, Markov chain Monte Carlo (MCMC) algorithms allow us to generate sets of samples that are employed to infer some relevant parameters of the underlying distributions. However, when the parameter space is high-dimensional, the performance of stochastic sampling algorithms is very sensitive to existing dependencies between parameters. In particular, this problem arises when one aims to sample from a high-dimensional Gaussian distribution whose covariance matrix does not present a simple structure. Another challenge is the design of Metropolis-Hastings proposals that make use of information about the local geometry of the target density in order to speed up the convergence and improve mixing properties in the parameter space, while not being too computationally expensive. These two contexts are mainly related to the presence of two heterogeneous sources of dependencies stemming either from the prior or the likelihood in the sense that the related covariance matrices cannot be diagonalized in the same basis. In this work, we address these two issues. Our contribution consists of adding auxiliary variables to the model in order to dissociate the two sources of dependencies. In the new augmented space, only one source of correlation remains directly related to the target parameters, the other sources of correlations being captured by the auxiliary variables. Experiments are conducted on two practical image restoration problems-namely the recovery of multichannel blurred images embedded in Gaussian noise and the recovery of signal corrupted by a mixed Gaussian noise. Experimental results indicate that adding the proposed auxiliary variables makes the sampling problem simpler since the new conditional distribution no longer contains highly heterogeneous correlations. Thus, the computational cost of each iteration of the Gibbs sampler is significantly reduced while ensuring good mixing properties.Entities:
Keywords: Bayesian methods; Gaussian models; MCMC; auxiliary variables; data augmentation; large scale problems
Year: 2018 PMID: 33265201 PMCID: PMC7512603 DOI: 10.3390/e20020110
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
Different alternatives for adding auxiliary variables.
| Problem | Proposed Auxiliary Variable | Resulting Conditional Density |
|---|---|---|
Restoration results. SNR: signal-to-noise ratio; BSNR: blurred SNR; PSNR: peak SNR; MMSE: minimum mean square error; SSIM: structural similarity.
| Initial | BSNR | 24.27 | 30.28 | 31.73 | 28.92 | 26.93 | 22.97 | 27.52 |
| PSNR | 25.47 | 21.18 | 19.79 | 22.36 | 23.01 | 26.93 | 23.12 | |
| SNR | 11.65 | 13.23 | 13.32 | 13.06 | 11.81 | 11.77 | 12.47 | |
| SSIM | 0.6203 | 0.5697 | 0.5692 | 0.5844 | 0.5558 | 0.6256 | 0.5875 | |
| MMSE | BSNR | 32.04 | 38.33 | 39.21 | 38.33 | 35.15 | 34.28 | 36.22 |
| PSNR | 28.63 | 25.39 | 23.98 | 26.90 | 27.25 | 31.47 | 27.27 | |
| SNR | 14.82 | 17.50 | 17.60 | 17.66 | 16.12 | 16.38 | 16.68 | |
| SSIM | 0.7756 | 0.8226 | 0.8156 | 0.8367 | 0.8210 | 0.8632 | 0.8225 | |
Figure 1From top to bottom: Original images–Degraded images–Restored images. (a) b = 2; (b) b = 4; (c) b = 6; (d) b = 2; (e) b = 4; (f) b = 6; (g) b = 2; (h) b = 4; (i) b = 6.
Figure 2Trace plot of the scale parameter in subband as time (horizontal subband in the first level of decomposition) with (dashed lines) and without (continuous line) auxiliary variables MALA: Metropolis-adapted Langevin algorithm; RW: random walk.
Mean and variance estimates of hyperparameters.
| RW | MALA | ||
|---|---|---|---|
| Mean | 0.67 | 0.67 | |
| Std. | (1.63 × | (1.29 × | |
| Mean | 0.83 | 0.83 | |
| Std. | (1.92 × | (2.39 × | |
| Mean | 0.62 | 0.61 | |
| Std. | (1.33 × | (1.23 × | |
| Mean | 0.24 | 0.24 | |
| Std. | (1.30 × | (1.39 × | |
| Mean | 0.37 | 0.37 | |
| Std. | (2.10 × | (2.42 × | |
| Mean | 0.21 | 0.21 | |
| Std. | (1.19 × | (1.25 × | |
| Mean | 0.08 | 0.08 | |
| Std. | (0.91 × | (1.08 × | |
| Mean | 0.13 | 0.13 | |
| Std. | (1.60 × | (1.64 × | |
| Mean | 0.07 | 0.07 | |
| Std. | (0.83 × | (1 × | |
| Mean | 7.80 × | 7.87 × | |
| Std. | (1.34 × | (2.12 × | |
| Mean | 1.89 × | 2.10 × | |
| Std. | (9.96 × | (2.24 × | |
Figure 3Hierarchical model for image deblurring under two-term mixed Gaussian noise.
Figure 4Visual results. From top to bottom: Original images–Degraded images—Restored images. (a) ; (b) ; (c) ; (d) : SNR dB, , , h: Gaussian std. 4; (e) : SNR = dB, , , , h: Uniform ; (f) : SNR dB, , , h: Gaussian std. 1.8; (g) : SNR dB, , , × ; (h) : SNR dB, , , × ; (i) : SNR dB, , , × .
Figure 5Chains of versus iteration/time.
Figure 6Chains of versus iteration/time.
Figure 7Chains of versus iteration/time.
Figure 8Chains of versus iteration/time.
Mean and variance estimates. RJPO: Reversible Jump Perturbation Optimization.
| RJPO | AuxV1 | AuxV2 | ||
|---|---|---|---|---|
| Mean | 4.78 × | 4.84 × | 4.90 × | |
| Std. | (1.39 × | (1.25 × | (9.01 × | |
| Mean | 12.97 | 12.98 | 12.98 | |
| Std. | (4.49 × | (4.82 × | (4.91 × | |
| Mean | 39.78 | 39.77 | 39.80 | |
| Std. | (0.13) | (0.14) | (0.13) | |
| Mean | 0.35 | 0.35 | 0.35 | |
| Std. | (2.40 × | (2.71 × | (2.72 × | |
| Mean | 143.44 | 143.19 | 145.91 | |
| Std. | (10.72) | (11.29) | (9.92) | |
Mixing results for the different proposed algorithms. First row: Time per iteration. Second row: Estimates of the mean square jump in stationarity. Third row: Estimates of the mean square jump per second in stationarity. Fourth row: Relative efficiency to RJPO.
| RJPO | AuxV1 | AuxV2 | |
|---|---|---|---|
| 5.27 | 0.13 | 0.12 | |
| 15.41 | 14.83 | 4.84 | |
| 2.92 | 114.07 | 40.33 | |
| Efficiency | 1 | 39 | 13.79 |