| Literature DB >> 35591118 |
Guanlei Gao1, Jie Cao1,2, Chun Bao1,2, Qun Hao1,2, Aoqi Ma1, Gang Li3.
Abstract
Image dehazing is challenging due to the problem of ill-posed parameter estimation. Numerous prior-based and learning-based methods have achieved great success. However, most learning-based methods use the changes and connections between scale and depth in convolutional neural networks for feature extraction. Although the performance is greatly improved compared with the prior-based methods, the performance in extracting detailed information is inferior. In this paper, we proposed an image dehazing model built with a convolutional neural network and Transformer, called Transformer for image dehazing (TID). First, we propose a Transformer-based channel attention module (TCAM), using a spatial attention module as its supplement. These two modules form an attention module that enhances channel and spatial features. Second, we use a multiscale parallel residual network as the backbone, which can extract feature information of different scales to achieve feature fusion. We experimented on the RESIDE dataset, and then conducted extensive comparisons and ablation studies with state-of-the-art methods. Experimental results show that our proposed method effectively improves the quality of the restored image, and it is also better than the existing attention modules in performance.Entities:
Keywords: Transformer; convolutional neural network; image dehazing
Mesh:
Year: 2022 PMID: 35591118 PMCID: PMC9105677 DOI: 10.3390/s22093428
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1Architecture of the proposed Transformer for image dehazing (TID).
Figure 2Architecture of the proposed Transformer-based channel attention module (TCAM).
Figure 3Architecture of the proposed attention module.
Parameter of the convolution layer.
| Layer | Kernel_Size/Padding | Output Channel |
|---|---|---|
| Conv1_1 | 1 × 1/0 | 3 |
| Conv1_2 | 3 × 3/1 | 3 |
| Conv1_3 | 5 × 5/2 | 9 |
| Conv2_1 | 3 × 3/1 | 3 |
| Conv2_2 | 5 × 5/2 | 3 |
| Conv2_3 | 7 × 7/3 | 9 |
| Conv3 | 3 × 3/1 | 3 |
Average PSNR/SSIM of dehazed results on the SOTS and HSTS dataset.
| Dataset | Fattal’s | FVR | DehazeNet | AOD-Net | EPDN | AECR-Net | Ours | |
|---|---|---|---|---|---|---|---|---|
| SOTS | PSNR | 161143 | 16.8931 | 18.7453 | 18.5211 | 20.1722 | 20.5466 | 21.4393 |
| SSIM | 0.7261 | 0.7484 | 0.8314 | 0.8314 | 0.8576 | 0.8642 | 0.8851 | |
| HSTS | PSNR | 17.7348 | 18.0142 | 21.2218 | 21.2218 | 22.3145 | 22.7693 | 23.8276 |
| SSIM | 0.8123 | 0.8217 | 0.8687 | 0.8687 | 0.8809 | 0.8914 | 0.9022 |
Figure 4Dehazing results on HSTS dataset. (a) Hazy image; (b) Fattal’s [26]; (c) FVR [43]; (d) DehazeNet [32]; (e) AOD-Net [15]; (f) EPDN [35]; (g) AECR-Net [36]; (h) TID (ours); (i) ground truth.
Figure 5Dehazing results on HSTS dataset. (a) Hazy image; (b) Fattal’s [26]; (c) FVR [43]; (d) DehazeNet [32]; (e) AOD-Net [15]; (f) EPDN [26]; (g) AECR-Net [36]; (h) TID (ours).
Figure 6Effects of ablation study (1).
Average PSNR/SSIM of ablation study (1).
| Non-Attention Module | Attention Module | |
|---|---|---|
| PSNR | 20.0566 | 21.4394 |
| SSIM | 0.8543 | 0.8851 |
Figure 7Effects of ablation study (2).
Average PSNR/SSIM of ablation study (2).
| SE-Net | CBAM | Ours | |
|---|---|---|---|
| PSNR | 20.2163 | 20.7560 | 21.4394 |
| SSIM | 0.8610 | 0.8593 | 0.8851 |
Figure 8Effects of ablation study (3).
Average PSNR/SSIM of ablation study (3).
| Non-Attention Module | Attention Module | |
|---|---|---|
| PSNR | 21.0335 | 21.4394 |
| SSIM | 0.8688 | 0.8851 |