| Literature DB >> 35271033 |
Wenxiang Chen1,2, Yingna Li1,2, Zhengang Zhao1,2.
Abstract
The vibration dampers can eliminate the galloping phenomenon of transmission lines caused by the wind. The detection of vibration dampers based on visual technology is an important issue. Current CNN-based methods struggle to meet the requirements of real-time detection. Therefore, the current vibration damper detection work has mainly been carried out manually. In view of the above situation, we propose a vibration damper detection-image generation model called DamperGAN based on multi-granularity Conditional Generative Adversarial Nets. DamperGAN first generates a low-resolution detection result image based on a coarse-grained module, then uses Monte Carlo search to mine the latent information in the low-resolution image, and finally injects this information into a fine-grained module through an attention mechanism to output high-resolution images and penalize poor intermediate information. At the same time, we propose a multi-level discriminator based on the multi-task learning mechanism to improve the discriminator's discriminative ability and promote the generator to output better images. Finally, experiments on the self-built DamperGenSet dataset show that the images generated by our model are superior to the current mainstream baselines in both resolution and quality.Entities:
Keywords: Monte Carlo search (MCS); conditional generative adversarial nets (CGAN); power transmission lines; unmanned aerial vehicle (UAV); vibration dampers detection
Year: 2022 PMID: 35271033 PMCID: PMC8914797 DOI: 10.3390/s22051886
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Test examples of our model on DamperGenSet.
Figure 2The framework of DamperGAN.
Figure 3The framework of the generator containing three parts: the global module, the Monte Carlo search, and the local module. First, we used the global module to obtain the LR image; then, we used the Monte Carlo for detail mining; and finally, we used the local module to obtain the HR image.
Figure 4The discriminator consists of three parts with the same structure: D1, D2, and D3. First, we used the shared convolutional layer to obtain the feature map of input; then, the feature map was down-sampled for 2 times and 4 times and output to D1, D2, and D3; finally, we obtained the discriminator scores.
The architecture of the generator.
| Network | Layer Information | Input | Output |
|---|---|---|---|
| Down-Sample | CONV-(N64, K7 × 7, S1, P3), IN, ReLU | (512, 512, 3) | (512, 512, 64) |
| CONV-(N128, K3 × 3, S2, P1), IN, ReLU | (512, 512, 64) | (256, 256, 128) | |
| CONV-(N256, K3 × 3, S2, P1), IN, ReLU | (256, 256, 128) | (128, 128, 256) | |
| CONV-(N512, K3 × 3, S2, P1), IN, ReLU | (128, 128, 256) | (64, 64, 512) | |
| CONV-(N1024, K3 × 3, S2, P1), IN, ReLU | (64, 64, 512) | (32, 32, 1024) | |
| Residual Block | CONV-(N1024, K3 × 3, S1, P1), IN, ReLU | (32, 32, 1024) | (32, 32, 1024) |
| CONV-(N1024, K3 × 3, S1, P1), IN, ReLU | (32, 32, 1024) | (32, 32, 1024) | |
| CONV-(N1024, K3 × 3, S1, P1), IN, ReLU | (32, 32, 1024) | (32, 32, 1024) | |
| CONV-(N1024, K3 × 3, S1, P1), IN, ReLU | (32, 32, 1024) | (32, 32, 1024) | |
| CONV-(N1024, K3 × 3, S1, P1), IN, ReLU | (32, 32, 1024) | (32, 32, 1024) | |
| CONV-(N1024, K3 × 3, S1, P1), IN, ReLU | (32, 32, 1024) | (32, 32, 1024) | |
| CONV-(N1024, K3 × 3, S1, P1), IN, ReLU | (32, 32, 1024) | (32, 32, 1024) | |
| CONV-(N1024, K3 × 3, S1, P1), IN, ReLU | (32, 32, 1024) | (32, 32, 1024) | |
| CONV-(N1024, K3 × 3, S1, P1), IN, ReLU | (32, 32, 1024) | (32, 32, 1024) | |
| Up-Sample | CONV-(N512, K3 × 3, S0.5, P1), IN, ReLU | (32, 32, 1024) | (64, 64, 512) |
| CONV-(N256, K3 × 3, S0.5, P1), IN, ReLU | (64, 64, 512) | (128, 128, 256) | |
| CONV-(N128, K3 × 3, S0.5, P1), IN, ReLU | (128, 128, 256) | (256, 256, 128) | |
| CONV-(N64, K3 × 3, S0.5, P1), IN, ReLU | (256, 256, 128) | (512, 512, 64) | |
| CONV-(N3, K7 × 7, S1, P3), IN, ReLU | (512, 512, 64) | (512, 512, 3) |
The architecture of the discriminators.
| Network | Layer Information | Input | Output |
|---|---|---|---|
| Input Layer | CONV-(N64, K4 × 4, S2, P2), Leaky ReLU | (512, 512, 3) | (256, 256, 64) |
| CONV-(N128, K4 × 4, S2, P2), IN, ReLU | (256, 256, 64) | (128, 128, 128) | |
| CONV-(N256, K4 × 4, S2, P2), IN, ReLU | (128, 128, 128) | (64, 64, 256) | |
| CONV-(N512, K4 × 4, S2, P2), IN, ReLU | (64, 64, 256) | (32, 32, 512) |
Figure 5Test examples of each model on the DamperGenSet dataset.
IS and FID of the different models.
| Model | InsuGenSet | |
|---|---|---|
| IS | FID | |
| Pix2Pix | 3.25 | 57.45 |
| CRN | 3.04 | 57.68 |
| X-Fork | 3.37 | 56.90 |
| X-Seq | 3.61 | 56.42 |
| SelectionGAN | 3.75 | 56.08 |
| DamperGAN | 3.83 | 55.31 |
SSIM, PSNR, SD, and FPS of the different models.
| Model | InsuGenSet | FPS | ||
|---|---|---|---|---|
| SSIM | PSNR | SD | ||
| Pix2Pix | 0.29 | 15.91 | 17.41 | 160 |
| CRN | 0.27 | 15.53 | 17.12 | 187 |
| X-Fork | 0.38 | 16.37 | 18.21 | 85 |
| X-Seq | 0.45 | 17.34 | 18.58 | 72 |
| SelectionGAN | 0.63 | 26.83 | 20.61 | 66 |
| DamperGAN | 0.70 | 28.14 | 22.13 | 63 |
Comparison of the effectiveness of the generator networks.
| IS | FID | FPS | |
|---|---|---|---|
| Single generator | 3.28 | 56.84 | 82 |
| Two-stage generator | 3.83 | 55.31 | 63 |
| Three-stage generator | 4.25 | 54.96 | 37 |
Introducing the Monte Carlo search time comparison.
| MCS | SSIM | PSNR | SD | FPS |
|---|---|---|---|---|
| Not introduced | 0.57 | 26.28 | 19.30 | 75 |
| 0.63 | 26.84 | 20.46 | 70 | |
| 0.68 | 27.60 | 21.37 | 67 | |
| 0.70 | 28.14 | 22.13 | 63 | |
| 0.72 | 28.47 | 22.54 | 58 | |
| 0.73 | 28.62 | 23.02 | 51 |
Comparison of the effectiveness of the discriminant networks.
| SSIM | PSNR | SD | FPS | |
|---|---|---|---|---|
| Single discriminator | 0.60 | 26.48 | 19.84 | 71 |
| Two-level discriminator | 0.65 | 27.26 | 20.62 | 67 |
| Three-level discriminator | 0.70 | 28.14 | 22.13 | 63 |
| Four-level discriminator | 0.72 | 28.53 | 22.79 | 58 |
The effect of different epoch numbers on the experimental results.
| Number of Epochs | SSIM | PSNR | SD | FPS |
|---|---|---|---|---|
| 50 | 0.32 | 16.72 | 17.84 | 65 |
| 100 | 0.58 | 18.56 | 19.05 | 64 |
| 150 | 0.64 | 24.47 | 21.31 | 63 |
| 200 | 0.70 | 28.14 | 22.13 | 63 |
| 250 | 0.68 | 27.92 | 21.86 | 61 |
Minimum training data experimental results.
| The Amount of Training Set | SSIM | PSNR | SD |
|---|---|---|---|
| 2500 (100%) | 0.70 | 28.14 | 22.13 |
| 2250 (90%) | 0.68 | 27.82 | 21.94 |
| 2000 (80%) | 0.65 | 25.86 | 20.25 |
| 1750 (70%) | 0.62 | 25.15 | 19.93 |
| 1500 (60%) | 0.56 | 23.83 | 17.42 |
Network parameters (Param.) and training time of the different models.
| Model | Param. | Training Time (h) |
|---|---|---|
| Pix2Pix | 47 M | 14.92 |
| CRN | 36 M | 10.88 |
| X-Fork | 62 M | 16.30 |
| X-Seq | 70 M | 18.57 |
| SelectionGAN | 78 M | 20.06 |
| DamperGAN | 82 M | 22.68 |