| Literature DB >> 35336291 |
Ruiming Jia1, Xin Chen1, Tong Li1, Jiali Cui1.
Abstract
Infrared image simulation is challenging because it is complex to model. To estimate the corresponding infrared image directly from the visible light image, we propose a three-level refined light-weight generative adversarial network with cascaded guidance (V2T-GAN), which can improve the accuracy of the infrared simulation image. V2T-GAN is guided by cascading auxiliary tasks and auxiliary information: the first-level adversarial network uses semantic segmentation as an auxiliary task, focusing on the structural information of the infrared image; the second-level adversarial network uses the grayscale inverted visible image as the auxiliary task to supplement the texture details of the infrared image; the third-level network obtains a sharp and accurate edge by adding auxiliary information of the edge image and a displacement network. Experiments on the public dataset Multispectral Pedestrian Dataset demonstrate that the structure and texture features of the infrared simulation image obtained by V2T-GAN are correct, and outperform the state-of-the-art methods in objective metrics and subjective visualization effects.Entities:
Keywords: generative adversarial network; image domain translation; infrared image simulation
Year: 2022 PMID: 35336291 PMCID: PMC8949294 DOI: 10.3390/s22062119
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1V2T-GAN network structure.
Figure 2The proposed network of G2.
Figure 3The structure of MFM.
Figure 4The proposed network of G.
Figure 5The proposed network of G.
Comparison of the algorithms in objective metrics.
| Methods | The Lower, The Better | The Higher, The Better | ||||
|---|---|---|---|---|---|---|
| Abs Rel | Avg log10 | RMS | PSNR | SSIM | ||
| Pix2pix [ | 0.248 | 0.107 | 0.906 | 0.571 | 22.431 | 0.985 |
| X-Fork [ | 0.314 | 0.130 | 1.074 | 0.480 | 20.692 | 0.984 |
| Selection-GAN [ | 0.284 | 0.112 | 0.958 | 0.554 | 21.976 | 0.982 |
| SEAN [ | 0.293 | 0.114 | 0.966 | 0.564 | 21.804 | 0.983 |
| LG-GAN [ | 0.262 | 0.102 | 0.886 | 0.616 | 22.601 | 0.989 |
| Ours | 0.247 | 0.099 | 0.850 | 0.623 | 22.908 | 0.990 |
Figure 6Compared with the computational efficiency of advanced algorithms.
Comparison of the different network structures.
| Network | The Lower, The Better | The Higher, The Better | ||||
|---|---|---|---|---|---|---|
| Rel | Avg log10 | RMS | PSNR | SSIM | ||
| One-level | 0.254 | 0.100 | 0.859 | 0.617 | 22.838 | 0.988 |
| Two-level | 0.254 | 0.099 | 0.853 | 0.619 | 22.872 | 0.990 |
| Three-level | 0.247 | 0.099 | 0.850 | 0.623 | 22.908 | 0.990 |
Figure 7Comparison of the one-level and two-level networks’ visualization results.
Figure 8Comparison of two-level and three-level networks’ visualization results.
Comparison of the different auxiliary tasks.
| Setup | The Lower, The Better | The Higher, The Better | ||||
|---|---|---|---|---|---|---|
| Rel | Avg log10 | RMS | PSNR | SSIM | ||
| −G | 0.257 | 0.103 | 0.876 | 0.609 | 22.674 | 0.989 |
| −G | 0.255 | 0.102 | 0.870 | 0.611 | 22.678 | 0.989 |
| −G | 0.249 | 0.101 | 0.855 | 0.615 | 22.811 | 0.990 |
| Ours | 0.247 | 0.099 | 0.850 | 0.623 | 22.908 | 0.990 |
Effectiveness of the edge auxiliary information.
| Methods | The Lower, The Better | The Higher, The Better | ||||
|---|---|---|---|---|---|---|
| Abs Rel | Avg log10 | RMS | PSNR | SSIM | ||
| − | 0.248 | 0.099 | 0.850 | 0.616 | 22.922 | 0.989 |
| Ours | 0.247 | 0.099 | 0.850 | 0.623 | 22.908 | 0.990 |
Comparison of the different lightweight convolutions.
| Methods | The Lower, The Better | The Higher, The Better | Params | ||||
|---|---|---|---|---|---|---|---|
| Abs Rel | Avg log10 | RMS | PSNR | SSIM | |||
| BSConv | 0.256 | 0.105 | 0.898 | 0.601 | 22.442 | 0.989 | 5.081M |
| DSConv | 0.259 | 0.108 | 0.916 | 0.593 | 22.257 | 0.989 | 5.126M |
| GhostModule | 0.260 | 0.105 | 0.891 | 0.604 | 22.565 | 0.987 | 15.263M |
| GConv | 0.247 | 0.099 | 0.850 | 0.623 | 22.908 | 0.990 | 15.235M |