| Literature DB >> 31547609 |
Yuxia Li1, Bo Peng2, Lei He3, Kunlong Fan4, Zhenxu Li5, Ling Tong6.
Abstract
Roads are vital components of infrastructure, the extraction of which has become a topic of significant interest in the field of remote sensing. Because deep learning has been a popular method in image processing and information extraction, researchers have paid more attention to extracting road using neural networks. This article proposes the improvement of neural networks to extract roads from Unmanned Aerial Vehicle (UAV) remote sensing images. D-Linknet was first considered for its high performance; however, the huge scale of the net reduced computational efficiency. With a focus on the low computational efficiency problem of the popular D-LinkNet, this article made some improvements: (1) Replace the initial block with a stem block. (2) Rebuild the entire network based on ResNet units with a new structure, allowing for the construction of an improved neural network D-Linknetplus. (3) Add a 1 × 1 convolution layer before DBlock to reduce the input feature maps, reducing parameters and improving computational efficiency. Add another 1 × 1 convolution layer after DBlock to recover the required number of output channels. Accordingly, another improved neural network B-D-LinknetPlus was built. Comparisons were performed between the neural nets, and the verification were made with the Massachusetts Roads Dataset. The results show improved neural networks are helpful in reducing the network size and developing the precision needed for road extraction.Entities:
Keywords: UAV sensors; convolutional neural net; image processing; road
Year: 2019 PMID: 31547609 PMCID: PMC6806354 DOI: 10.3390/s19194115
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Unmanned Aerial Vehicle (UAV) Remote Sensing Images.
Figure 2Mosaic Unmanned Aerial Vehicle images (Luoyang Region).
Figure 3Initial Software Interface and Annotation Results.
Figure 4Label Result Image.
Figure 5The results of spatial geometric transformation of the label (a) original label; (b) vertical flip; (c) horizontal flip; (d) main diagonal rotation; (e) sub diagonal transpose; (f) 30 degree clockwise rotation; (g) 60 degree clockwise rotation; (h) 90 degree clockwise rotation; (i) 180 degree clockwise rotation; (j) 270 degree clockwise rotation; (k,l) random shift and scaling (approximately −10–10%).
Figure 6The contrast of initial block and stem block: (a) initial block and (b) stem block.
Figure 7The structure of the different residual units: (a) Original residual unit and (b) New residual unit.
Figure 8The structure of DBlock.
Figure 9D-LinkNetPlus structure chart.
Figure 10The structure of DBlockPlus.
D-LinkNetPlus structure parameters.
| Layer Name | Output Size | D-LinkNetPlus_50 | D-LinkNetPlus_101 |
|---|---|---|---|
| Stem block |
| Conv | Conv |
| Encoder1 |
| Conv | Conv |
| Encoder2 |
| Conv | Conv |
| Encoder3 |
| Conv | Conv |
| Encoder4 |
| Conv | Conv |
| Center |
| DBlock | DBlock |
| Decoder4 |
|
|
|
| Decoder3 |
|
|
|
| Decoder2 |
|
|
|
| Decoder1 |
|
|
|
| F1 |
| Deconv | Deconv |
| Logits |
| Conv | Conv |
Figure 11Comparison results of simple scene experiment. (a) Original image; (b) label image; (c) D-LinkNet_50; (d) D-LinkNet_101; (e) D-LinkNetPlus_50; and (f) D-LinkNetPlus_101.
Figure 12Comparison results of general scene experiments. (a) Input image; (b) label image; (c) D-LinkNet_50; (d) D-LinkNet_101; (e) D-LinkNetPlus_50; and (f) D-LinkNetPlus_101.
Figure 13Comparison results of complex scene experiment. (a) Original image; (b) label image; (c) D-LinkNet_50; (d) D-LinkNet_101; (e) D-LinkNetPlus_50; and (f) D-LinkNetPlus_101.
Comparison results of network size and road precision between D-LinkNet and D-LinkNetPlus.
| Network Name | Network Size | IoU |
|---|---|---|
| D-LinkNet_50 | 792 M | 51.02% |
| D-LinkNet_101 | 0.98 G | 52.67% |
| D-LinkNetPlus_50 | 686 M | 51.85% |
| D-LinkNetPlus_101 | 758 M | 52.87% |
Figure 14Comparison results of simple scene experiment. (a) Input image; (b) label image (c) D-LinkNet_50; (d) D-LinkNet_101; (e) B-D-LinkNetPlus_50; and (f) B-D-LinkNetPlus_101.
Figure 15Comparison results of general scene experiments. (a) Input image; (b) label image; (c)D-LinkNet_50; (d) D-LinkNet_101; (e) B-D-LinkNetPlus_50; and (f) B-D-LinkNetPlus_101.
Figure 16Comparison results of complex scene experiment. (a) input image; (b) label image; (c) D-LinkNet_50; (d) D-LinkNet_101; (e) B-D-LinkNetPlus_50; and (f) B-D-LinkNetPlus_101.
D-LinkNet and B-D-LinkNetPlus network size and precision comparison results.
| Network Name | Network Model Size | IoU |
|---|---|---|
| D-LinkNet_50 | 792 M | 51.02% |
| D-LinkNet_101 | 0.98 G | 52.67% |
| B-D-LinkNetPlus_50 | 298 M | 52.86% |
| B-D-LinkNetPlus_101 | 370 M | 52.94% |
Comparison results of road precision between D-LinkNet, D-LinkNetPlus, and B-D-LinkNetPlus.
| Network Name | Network Model Size | IoU |
|---|---|---|
| D-LinkNet_50 | 702 M | 57.18% |
| D-LinkNet_101 | 758 M | 57.43% |
| D-LinkNetPlus_50 | 582 M | 57.58% |
| D-LinkNetPlus_101 | 636 M | 57.64% |
| B-D-LinkNetPlus_50 | 371 M | 59.29% |
| B-D-LinkNetPlus_101 | 434 M | 59.45% |
Figure 17Comparison results of complex scene experiment. (a) Input image; (b) label image; (c) D-LinkNet_50; and (d) D-LinkNet_101.
Figure 18Comparison results of complex scene experiment. (a) D-LinkNetPlus_50; (b) D-LinkNetPlus_101; (c) B-D-LinkNetPlus_50; and (d) B-D-LinkNetPlus_101.