| Literature DB >> 35684852 |
Hesen Feng1,2, Lihong Ma1,2, Jing Tian3.
Abstract
Image super-resolution aims to reconstruct a high-resolution image from its low-resolution counterparts. Conventional image super-resolution approaches share the same spatial convolution kernel for the whole image in the upscaling modules, which neglect the specificity of content information in different positions of the image. In view of this, this paper proposes a regularized pattern method to represent spatially variant structural features in an image and further exploits a dynamic convolution kernel generation method to match the regularized pattern and improve image reconstruction performance. To be more specific, first, the proposed approach extracts features from low-resolution images using a self-organizing feature mapping network to construct regularized patterns (RP), which describe different contents at different locations. Second, the meta-learning mechanism based on the regularized pattern predicts the weights of the convolution kernels that match the regularized pattern for each different location; therefore, it generates different upscaling functions for images with different content. Extensive experiments are conducted using the benchmark datasets Set5, Set14, B100, Urban100, and Manga109 to demonstrate that the proposed approach outperforms the state-of-the-art super-resolution approaches in terms of both PSNR and SSIM performance.Entities:
Keywords: RPB-RDN; dynamic convolution kernel; image super-resolution; multi-task learning; regularized pattern
Year: 2022 PMID: 35684852 PMCID: PMC9185547 DOI: 10.3390/s22114231
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1A conceptual illustration of our proposed SR method. Pixels at different positions are upsampled using different convolution kernels that match the regularized pattern of the current position.
Figure 2SR images generated by the proposed method at multiple scales.
Figure 3Visualization of dynamic convolution weights that are generated by our proposed approach. The first column presents the original test images, and the second column presents the visualized convolution weight variation values calculated using Equation (10). (a) 253027 from B100 dataset [19]. (b) img59 from Urban100 dataset [20]. (c) YumeiroCooking from Manga109 dataset [21].
Figure 4An overview of the network structure of the proposed SR approach.
PSNR (dB) and SSIM performance on proposed regularized pattern extraction network. The best performance is highlighted in the bold format.
| Methods | Metric | B100 [ | Urban100 [ | Manga109 [ | ||||||
|---|---|---|---|---|---|---|---|---|---|---|
| X2 | X3 | X4 | X2 | X3 | X4 | X2 | X3 | X4 | ||
| Baseline model | PSNR | 32.34 | 29.27 | 27.75 | 33.00 | 28.90 | 26.68 | 39.31 | 34.41 | 31.36 |
| RPB-RDN (Ours) | PSNR |
|
|
|
|
|
|
|
|
|
PSNR (dB) performance evaluation on the proposed convolution weight prediction method using the B100 dataset [19]. The best performance is highlighted in the bold format.
| Methods | X1.1 | X1.2 | X1.3 | X1.4 | X1.5 | X1.6 | X1.7 | X1.8 | X1.9 | X2.0 |
|---|---|---|---|---|---|---|---|---|---|---|
| Bicubic | 36.56 | 35.01 | 33.84 | 32.93 | 32.14 | 31.49 | 30.90 | 30.38 | 29.97 | 29.55 |
| Meta-RDN [ | 42.82 | 40.04 | 38.28 | 36.95 |
| 34.90 | 34.13 | 33.45 |
| 32.35 |
| Ours |
|
|
|
|
|
|
|
|
|
|
| Methods | X2.1 | X2.2 | X2.3 | X2.4 | X2.5 | X2.6 | X2.7 | X2.8 | X2.9 | X3.0 |
| Bicubic | 29.18 | 28.87 | 28.57 | 28.31 | 28.13 | 27.89 | 27.66 | 27.51 | 27.31 | 27.19 |
| Meta-RDN [ | 31.82 | 31.41 | 31.06 | 30.62 | 30.45 | 30.13 | 29.82 | 29.67 | 29.40 |
|
| Ours |
|
|
|
|
|
|
|
|
|
|
| Methods | X3.1 | X3.2 | X3.3 | X3.4 | X3.5 | X3.6 | X3.7 | X3.8 | X3.9 | X4.0 |
| Bicubic | 26.98 | 26.89 | 26.59 | 26.60 | 26.42 | 26.35 | 26.15 | 26.07 | 26.01 | 25.96 |
| Meta-RDN [ | 28.87 | 28.79 | 28.68 | 28.54 | 28.32 |
| 28.04 | 27.92 | 27.82 | 27.75 |
| Ours |
|
|
|
|
|
|
|
|
|
|
The running time in B100 of various SR approaches (ms). The best performance is highlighted in the bold format.
| Methods | X2 | X3 | X4 |
|---|---|---|---|
| RDN |
|
|
|
| Meta-RDN [ | 14.4 | 14.8 | 16.4 |
| LIIF-RDN [ | 21.3 | 22.7 | 24.9 |
| RPB-RDN (Ours) | 15.1 | 15.3 | 16.5 |
The PSNR (dB) and SSIM performance in texture dataset Texture. The best performance is highlighted in the bold format.
| Methods | Metric | X2 | X3 | X4 |
|---|---|---|---|---|
| Meta-RDN [ | PSNR(dB)/SSIM | 34.22/0.9312 | 30.30/0.8601 | 28.08/0.7976 |
| LIIF-RDN [ | 34.19/0.9309 | 30.31/0.8601 | 28.09/0.7975 | |
| RPB-RDN (Ours) |
|
|
|
The PSNR (dB) and SSIM performance comparison of various SR approaches. The best performance is highlighted in the bold format.
| Dataset | The PSNR (dB) Performance | |||||
|---|---|---|---|---|---|---|
| Scale Factor | Bicubic | RDN [ | Meta-RDN [ | LIIF-RDN [ | Ours | |
| Set5 [ | X2 | 33.66 | 38.22 | 38.17 | 38.23 | |
| Set14 [ | X2 | 30.24 | 34.01 | 34.04 | 33.97 | |
| B100 [ | X2 | 29.56 | 32.34 | 32.35 | 32.32 |
|
| Urban100 [ | X2 | 26.88 | 32.89 | 32.92 | 32.87 |
|
| Manga109 [ | X2 | 30.80 | 39.18 | 39.18 | 39.01 |
|
|
|
| |||||
|
|
|
| ||||
| Set5 [ | X2 | 0.9299 | 0.9611 | 0.9610 | 0.9611 | |
| Set14 [ | X2 | 0.8688 | 0.9212 | 0.9213 | 0.9208 |
|
| B100 [ | X2 | 0.8431 | 0.9017 |
| 0.9010 | 0.9014 |
| Urban100 [ | X2 | 0.8403 | 0.9353 | 0.9361 | 0.9350 |
|
| Manga109 [ | X2 | 0.9339 | 0.9780 | 0.9780 |
| |
Figure 5Qualitative performance comparison of various SR approaches.