| Literature DB >> 35808203 |
Sotiris Karavarsamis1, Ioanna Gkika1, Vasileios Gkitsas1, Konstantinos Konstantoudakis1, Dimitrios Zarpalas1.
Abstract
This survey article is concerned with the emergence of vision augmentation AI tools for enhancing the situational awareness of first responders (FRs) in rescue operations. More specifically, the article surveys three families of image restoration methods serving the purpose of vision augmentation under adverse weather conditions. These image restoration methods are: (a) deraining; (b) desnowing; (c) dehazing ones. The contribution of this article is a survey of the recent literature on these three problem families, focusing on the utilization of deep learning (DL) models and meeting the requirements of their application in rescue operations. A faceted taxonomy is introduced in past and recent literature including various DL architectures, loss functions and datasets. Although there are multiple surveys on recovering images degraded by natural phenomena, the literature lacks a comprehensive survey focused explicitly on assisting FRs. This paper aims to fill this gap by presenting existing methods in the literature, assessing their suitability for FR applications, and providing insights for future research directions.Entities:
Keywords: deep learning; deep neural networks; dehazing; deraining; desnowing
Mesh:
Year: 2022 PMID: 35808203 PMCID: PMC9269588 DOI: 10.3390/s22134707
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Listing of datasets used for the deraining task.
| Dataset | Synthetic (S)/Real (R) | Indoor (I)/Outdoor (O) | Pairs | Year |
|---|---|---|---|---|
| Rain12600
[ | S | O | 14,000 | 2017 |
| Rain12000 [ | S | O | 12,000 | 2018 |
| Rain1400 [ | S | O | 1400 | 2017 |
| Rain800 [ | R | O | 800 | 2020 |
| Rain12 [ | S | O | 12 | 2016 |
| Test100 [ | S | O | 100 | 2020 |
| Test1200 [ | S | O | 1200 | 2018 |
| RainTrainH [ | S | O | 1800 | 2018 |
| RainTrainL [ | S | O | 200 | 2018 |
| Rain100H [ | S | O | 100 | 2020 |
| Rain100L [ | S | O | 100 | 2020 |
| Rain200H [ | S | O | 2000 | 2017 |
| RID [ | R | O | 2495 | 2019 |
| RIS [ | R | O | 2048 | 2019 |
| DAWN/Rainy [ | R | O | 200 | 2020 |
| NTURain [ | S | O | 8 (videos) | 2018 |
| SPA-Data [ | R | O | 29,500 | 2019 |
Listing of datasets used for the desnowing task.
| Dataset | Synthetic (S)/Real (R) | Indoor (I)/Outdoor (O) | Pairs | Year |
|---|---|---|---|---|
| Snow-100K
[ | S & R | O | 100,000+ | 2018 |
| SRRS [ | S & R | I & O | 16,000 | 2020 |
| CSD [ | S | I & O | 10,000 | 2021 |
| SITD [ | S | O | 3000 | 2019 |
Listing of datasets used for the dehazing task.
| Dataset | Synthetic (S)/Real (R) | Indoor (I)/Outdoor (O) | Pairs | Year |
|---|---|---|---|---|
| FRIDA
[ | S | O | 72 | 2010 |
| FRIDA2 [ | S | O | 264 | 2012 |
| CHIC [ | G | I | 18 | 2016 |
| RESIDE [ | S & R | I & O | 10,000+ | 2018 |
| D-HAZY [ | S | I |
| 2016 |
| I-HAZE [ | G | I | 35 | 2018 |
| O-HAZE [ | G | O | 45 | 2018 |
| DENSE-HAZE [ | G | O | 35 | 2019 |
| NH-HAZE [ | G | O | 55 | 2020 |
| HazeRD [ | S | O | 70 | 2017 |
| BeDDE [ | R | O |
| 2020 |
| 4KID [ | S | O | 10,000 | 2021 |
| REVIDE [ | G | I | 2021 |
Listing of surveyed research papers.
| Category | Method | Model | Short Description | Year |
|---|---|---|---|---|
| CNN- | DetailNet [ | ACM | reduces mapping range; promotes HF details | 2017 |
| Residual-guide [ | ACM | cascaded; residuals; coarse-to-fine | 2018 | |
| NLEDN [ | ACM | end-to-end, non-locally-enhanced; spatial correlation | 2018 | |
| DID-MDN [ | ACM | density-aware multi-stream densely connected CNN | 2018 | |
| DualCNN [ | ACM | estimation of structures and details | 2018 | |
| Scale-free [ | HRMLL | wavelet analysis | 2019 | |
| DMTNet [ | ACM | symmetry reduces complexity; multidomain translation | 2021 | |
| UC-PFilt [ | ACM | predictive kernels; removes residual rain traces | 2022 | |
| SAPNet [ | N/A | PDUs; unsupervised background segmentation; perceptual loss | 2022 | |
| DDC [ | SBM | decomposition and composition network; rain mask | 2019 | |
| DerainNet [ | ACM | non-linear rainy-to-clear mapping | 2017 | |
| PCNet [ | MRSL | learns joint features of rainy and clear content | 2021 | |
| Spatial Attention [ | ACM | human supervision; global-to-local attention | 2019 | |
| memory encoder–decoder [ | ACM | encoder–decoder architecture with memory | 2022 | |
| Attention | APAN [ | ACM | multi-scale pyramid representation; attention | 2021 |
| IADN [ | ACM | self-similarity of rain; mixed attention mechanism; fusion | 2020 | |
| DECAN [ | ACM | detail-guided channel attention module identifies low-level features; background repair network | 2021 | |
| DAF-Net [ | DRM | end-to-end model; depth-attentional features learning | 2019 | |
| SIR [ | ACM | encoder–decoder embedding; layered LSTM | 2022 | |
| RadNet [ | ACM/ Raindrop | restores raindrops and rainstreaks; handles single-type, superimposed-type or blended-type data | 2021 | |
| DARGNet [ | HRM | dual-attention (spatial and channel) | 2021 | |
| task-adaptive attention [ | N/A | task-adaptive, task-channel, task-operation attention | 2022 | |
| Generative | DerainAttentionGAN [ | ACM | uses Cycle-GAN; attention | 2022 |
| DerainCycleGAN [ | ACM | CycleGAN transfer learning; unsupervised attention | 2021 | |
| RCA-cGAN [ | ACM | rain streak characteristics; integration with cGAN | 2022 | |
| RainGAN [ | Raindrop | raindrop removal as many-to-one translation | 2022 | |
| UD-GAN [ | ACM | GAN; self-supervised constraints from intrinsic statistics | 2019 | |
| HeavyRainStorer [ | HRM | 2-stage network; physics-based backbone; depth-guided GAN | 2019 | |
| ID-CGAN [ | ACM | conditional GAN with additional constraint | 2020 | |
| AttGAN [ | Raindrop | attentive GAN; learns rain structure | 2018 | |
| Multi-scale | MSPFN [ | N/A | streak correlations; multi-scale progressive fusion | 2020 |
| MRADN [ | ACM | multi-scale residual aggregation; multi-scale context aggregation; multi-resolution feature extraction | 2021 | |
| LFDN [ | N/A | encoder–decoder architecture; encoder with multi-scale analysis; decoder with feature distillation; module fusion | 2021 | |
| SphinxNet [ | N/A | AEs for maximum spatial awareness; convolutional layers; skip concatenation connections | 2021 | |
| DFPGN [ | ACM | cross-scale information merge; cross-layer feature fusion | 2021 | |
| GAGN [ | ACM | context-wise; multi-scale analysis | 2022 | |
| UMRL [ | ACM | UMRL network learns rain content at different scales | 2019 | |
| Different | JORDER [ | HRM | multi-task learning; priors on equation parameters | 2020 |
| FLUID [ | N/A | few-shot; self-supervised; inpainting | 2022 | |
| Semi-supervised CNN [ | ACM | adapts to unpaired data by training on paired data | 2019 | |
| Recurrent | PReNet [ | ACM | repeated ResNet; recursive; multi-scale info extraction | 2019 |
| recurrent residual multi-scale [ | MRSL | residual multi-scale pyramid; coarse-to-fine progressive rain removal; attention map; multi-scale kernel selection | 2022 | |
| Scale-aware [ | HRM | multiple subnetworks handle range of rain characteristics | 2017 | |
| RESCAN [ | Equation ( | contextual dilated network; squeeze-and-excitation block | 2018 | |
| Pyramid Derain [ | ACM | Gaussian–Laplacian pyramid decomposition | 2019 | |
| DRN [ | ACM | multi-stage residual network with two residual blocks | 2019 | |
| NCANet [ | Equation ( | rain streaks as residuals sum; recurrent | 2022 | |
| PRRNet [ | ACM | stereo; semantic segmentation; multi-view fusion | 2021 | |
| GTA-Net [ | ACM | multi-stream coarse; single-stream fine | 2021 |
Figure 1Schematic of the DerainCycleGAN method by Wei et al. [66]. The model follows the CycleGAN regime and employs URAD (Unsupervised Rain Attentive Detector) modules to attend to rain information and guide intermediate data projections.
Figure 2Schematic of the SIR method by Li et al. [78]. The model employs an autoencoder for single image deraining, regularization loss functions and deterministically generated features that encourage the autoencoder to attend to rain streak features.
Figure 3Schematic of the SphinxNet method by Jasuja et al. [84]. Encoder and decoder modules are arranged hierarchically to allow for multi-scale feature encoding and decoding.
Figure 4Schematic of the method by Zhang et al. [85]. The (a) shows the MSKSN model, and (b) shows the MKSB module.
Listing of references on the desnowing problem.
| Category | Method | Short Description | Year |
|---|---|---|---|
| CNN- | HDCW-Net [ | DTCWT analysis; recursively computes HF component; neural network | 2021 |
| Generative | cGAN [ | separates the background from snowy regions; uses compositional loss | 2019 |
| JSTASR [ | handles transparent/non-transparent snow particles; modified partial | 2020 | |
| DesnowGAN [ | DNN with top-down pathway and lateral cross-resolution connections; | 2020 | |
| Multi-scale | DesnowNet [ | accurately corrects image content by estimating and restoring details in | 2018 |
| MS-SDN [ | multi-scale convolutional subnetwork extracts feature maps; stacked | 2019 | |
| DDMSNet [ | multi-scale representation from pixel-level and feature-level input; | 2021 |
Figure 5Schematic of the method by Li et al. [106]. A GAN-based architecture learns to extract the clean background image component and the rain streak layer component from an input rainy image by means of the composition loss function.
Figure 6Schematic of the JSTASR method by Chen et al. [32]. The model comprises three modules. First, it contains a joint size and transparency-aware snow removal module. Secondly, it employs a veiling effect recovery module, and finally it comprises a size-aware clean image discriminator.
Figure 7Schematic of the DesnowGAN method by Jaw et al. [107]. A GAN-based snow removal module, a refinement module and a discriminator module guide single image desnowing.
Figure 8Schematic of the DesnowNet method by Liu et al. [29].
Listing of the surveyed research papers and articles on the dehazing problem.
| Category | Method | Short Description | Year |
|---|---|---|---|
| CNN- | Dehazenet [ | 3-layer CNN, BReLU activation function | 2016 |
| AOD-Net [ | lightweight, transformed ASM | 2017 | |
| Light-DehazeNet [ | lightweight, transformed ASM, CVR module | 2021 | |
| FFA-Net [ | attention-based feature fusion structure | 2020 | |
| AECR-Net [ | AE-like, contrastive learning, feature fusion | 2021 | |
| Multi-scale | MSFFA-Net [ | multi-scale grid network, feature fusion | 2021 |
| GDNet [ | 3 sub-processes, multi-scale grid network | 2019 | |
| MSCNN [ | 2 nets: coarse- and fine-scale | 2016 | |
| MSCNN-HE [ | 3 nets: coarse-, fine-scale and holistic edge guided | 2020 | |
| EMRA-Net [ | 2 nets: TRA-CNN and EA-CNN | 2021 | |
| MSBDN [ | dense feature fusion module, boosted decoder | 2020 | |
| FAMED-Net [ | 3 encoders at different scales, fusion module | 2019 | |
| PGC [ | PGC and DRB blocks | 2020 | |
| MSRA-Net [ | CIELAB, 2 subnets (luminance, chrominance) | 2022 | |
| MSDFN [ | depth-aware dehazing | 2021 | |
| DMPHN [ | non-homogeneous haze, multi-patch architecture | 2020 | |
| TDN [ | 3 subnets: coarse-, fine-scale and haze density | 2020 | |
| Jo et al. [ | selective residual blocks | 2021 | |
| Generative | DCPDN [ | generator with 2 subnets, edge-preserving loss function | 2018 |
| DehazeGAN [ | ASM-based GAN | 2018 | |
| DDN [ | ASM-based, unpaired supervision | 2018 | |
| GFN [ | fusion based, employs a hazy image and 3 derived inputs | 2018 | |
| EPDN [ | multi-resolution generator, multi-scale discriminator, enhancer | 2019 | |
| cGAN [ | cGAN with encoder–decoder architecture | 2018 | |
| Kan et al. [ | cGAN, UR-Net as a generator, flexibility in image size | 2022 | |
| Cycle-Dehaze [ | CycleGan based, unpaired supervision | 2018 | |
| CDNet [ | CycleGan based, encoder–decoder architecture for the generator | 2019 | |
| Cycle-Defog2Refog [ | 2 transformation paths with 2-stage mapping strategy in each | 2020 | |
| UCDN [ | CycleGan based with a conditional disentangle network | 2020 | |
| DCA-CycleGAN [ | generator with 2 subnets, 4 discriminators | 2022 | |
| Park et al. [ | fusion of cGAN and CycleGAN | 2020 | |
| FD-GAN [ | integration of HF and LF information in the discriminator | 2020 | |
| DW-GAN [ | generator with a DWT and a Knowledge Adaptation Branch | 2021 | |
| TMS-GAN [ | 2 subnets: a haze-generation and a haze-removal GAN | 2021 | |
| RL-based | Dehaze-RL [ | actions: 11 dehazing algorithms, reward function: PSNR and SSIM | 2020 |
| DDRL [ | depth-aware dehazing | 2020 | |
| Knowledge | KDDN [ | teacher-student (dehazing) net | 2020 |
| Shao et al. [ | domain adaptation using a bidirectional translation net | 2020 | |
| PSD [ | domain adaptation by unsupervised fine-tuning (real domain) a | 2021 | |
| Yu et al. [ | 2-branch net: transfer learning and current data fitting subnets | 2021 | |
| Unsupervised/ | Golts et al. [ | unsupervised, DCP loss | 2019 |
| Li et al. [ | 2-branch: supervised and unsupervised subnets | 2019 | |
| RefineDNet [ | 2-stage network: DCP and adversarial learning stages | 2021 | |
| YOLY [ | self-supervised, 3 joint disentanglement subnetworks | 2021 |
Figure 9The architecture of Trident Dehazing Network [134]. ⊕ refers to tensor addition and ⊗ to tensor multiplication.
Figure 10The architecture of the Cycle-Dehaze Model [143]. G and F are the generators, and D_x and D_y the discriminators.
Quantitative results of deraining methods along with their running time. ↑ means higher is better. In bold we indicate the best result for each dataset.
| Dataset | Method | PSNR ↑ | SSIM ↑ | FPS ↑ | Image Resolution | Classification |
|---|---|---|---|---|---|---|
| Test1200 | RESCAN
[ | 30.51 | 0.882 | 1.83 |
| non-real-time |
| MSPFN [ | 32.39 |
| 1.97 |
| non-real-time | |
| PReNet [ | 31.36 | 0.911 | 6.13 |
| near-real-time | |
| IADN [ |
|
| 7.57 |
| near-real-time | |
| DDC [ | 28.65 | 0.854 | 8.00 |
| near-real-time | |
| DerainNet [ | 23.38 | 0.835 | 13.51 |
| near-real-time | |
| PCNet [ | 32.03 | 0.913 | 16.12 |
| near-real-time | |
| UMRL [ | 21.15 | 0.770 | 20.00 |
| real-time | |
| PCNet-fast [ | 31.45 | 0.906 | 35.71 |
| real-time | |
| LPNET [ | 25.00 | 0.782 |
|
| real-time | |
| Rain100L | JORDER [ |
| 0.921 | 5.55 |
| near-real-time |
| DDN [ | 31.12 | 0.926 | 6.25 |
| near-real-time | |
| ResGuideNet3 [ | 30.79 |
|
|
| near-real-time |
Quantitative results of desnowing methods along with their running time. ↑ means higher is better. In bold we indicate the best result among the evaluated methods.
| Dataset | Method | PSNR ↑ | SSIM ↑ | FPS ↑ | Image Resolution | Classification |
|---|---|---|---|---|---|---|
| Snow-100K | DesnowNet
[ |
| 0.930 | 0.72 |
| non-real-time |
| MS-SDN [ | 29.25 |
| 2.38 |
| non-real-time | |
| JSTASR [ | 28.61 | 0.864 | 2.77 |
| non-real-time | |
| DesnowGAN [ | 28.18 | 0.912 |
|
| real-time |
Quantitative results of dehazing methods along with their running time. ↑ means higher is better. In bold we indicate the best result among the evaluated methods.
| Dataset | Method | PSNR ↑ | SSIM ↑ | FPS ↑ | Image Resolution | Classification |
|---|---|---|---|---|---|---|
| SOTS | FFA-Net
[ | 36.39 | 0.988 | 0.57 |
| non-real-time |
| Li et al. [ | 24.44 | 0.890 | 0.89 |
| non-real-time | |
| MSCNN-HE [ | 21.56 | 0.860 | 1.20 |
| non-real-time | |
| TDN [ | 34.59 | 0.975 | 1.58 |
| near-real-time | |
| DW-GAN [ | 35.94 | 0.986 | 2.08 |
| near-real-time | |
| Light-DehazeNet [ | 28.39 | 0.948 | 2.38 |
| non-real-time | |
| PGC [ | 28.78 | 0.956 | 3.17 |
| near-real-time | |
| MSFFA-Net [ |
| 0.990 | 3.23 |
| near-real-time | |
| DehazeNet [ | 21.14 | 0.847 | 3.33 |
| near-real-time | |
| EPDN [ | 25.06 | 0.923 | 3.41 |
| near-real-time | |
| Golts et al. [ | 24.08 | 0.933 | 3.57 |
| near-real-time | |
| GDNet [ | 32.16 | 0.983 | 3.60 |
| near-real-time | |
| MSCNN [ | 17.57 | 0.810 | 3.85 |
| near-real-time | |
| YOLY [ | 19.41 | 0.832 | 4.76 |
| near-real-time | |
| Yu et al. [ | 36.61 |
| 11.24 |
| real-time | |
| cGAN [ | 26.63 | 0.942 | 19.23 |
| real-time | |
| GFN [ | 22.30 | 0.880 | 20.40 |
| real-time | |
| DCPDN [ | 19.39 | 0.650 | 23.98 |
| real-time | |
| FD-GAN [ | 23.15 | 0.920 | 65.00 |
| real-time | |
| DMPHN [ | 16.94 | 0.617 | 68.96 |
| real-time | |
| FAMED-Net [ | 25.00 | 0.917 | 86.20 |
| real-time | |
| AOD-Net [ | 19.06 | 0.850 |
|
| real-time |
Listing of loss functions used by desnowing methods.
| Loss Function | Ref. | Description |
|---|---|---|
|
| Chen et al. [ | Complex Wavelet Loss |
|
| Chen et al. [ | Contradict Channel Loss |
|
| Chen et al. [ | Perceptual Loss |
|
| Liu et al. [ | Lightweight Pyramid Loss |
|
| Liu et al. [ | Total Variation Loss |
|
| Liu et al. [ | Style Loss |