| Literature DB >> 30998770 |
Xu Ma1, Xiangwu Deng1, Long Qi1, Yu Jiang1, Hongwei Li1, Yuwei Wang1, Xupo Xing1.
Abstract
To reduce the cost of production and the pollution of the environment that is due to the overapplication of herbicide in paddy fields, the location information of rice seedlings and weeds must be detected in site-specific weed management (SSWM). With the development of deep learning, a semantic segmentation method with the SegNet that is based on fully convolutional network (FCN) was proposed. In this paper, RGB color images of seedling rice were captured in paddy field, and ground truth (GT) images were obtained by manually labeled the pixels in the RGB images with three separate categories, namely, rice seedlings, background, and weeds. The class weight coefficients were calculated to solve the problem of the unbalance of the number of the classification category. GT images and RGB images were used for data training and data testing. Eighty percent of the samples were randomly selected as the training dataset and 20% of samples were used as the test dataset. The proposed method was compared with a classical semantic segmentation model, namely, FCN, and U-Net models. The average accuracy rate of the SegNet method was 92.7%, whereas the average accuracy rates of the FCN and U-Net methods were 89.5% and 70.8%, respectively. The proposed SegNet method realized higher classification accuracy and could effectively classify the pixels of rice seedlings, background, and weeds in the paddy field images and acquire the positions of their regions.Entities:
Mesh:
Year: 2019 PMID: 30998770 PMCID: PMC6472823 DOI: 10.1371/journal.pone.0215676
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Fig 1Rice seedlings and weeds images in the paddy field.
Fig 2Image-label example.
(a) original image and (b) the corresponding GT labels.
Number of pixels with classes and the class weight coefficients.
| Pixel type | Percentage/% | Class weight coefficients |
|---|---|---|
| Weed | 5.028 | 2.280 |
| Rice seedling | 11.517 | 1.000 |
| Background | 83.455 | 0.138 |
Fig 3Network architecture of SegNet.
Encoder and decoder parameters of SegNet.
| Encoder | Layer type | Feature size | Number of features | Decoder | Layer type | Feature size | Number of features |
|---|---|---|---|---|---|---|---|
| Conv1_1 | 3×3 | 64 | Conv5_3 | 3×3 | 512 | ||
| Conv1_2 | 3×3 | 64 | Conv5_2 | 3×3 | 512 | ||
| Conv2_1 | 3×3 | 128 | Conv5_1 | 3×3 | 512 | ||
| Conv2_2 | 3×3 | 128 | Conv4_3 | 3×3 | 512 | ||
| Conv3_1 | 3×3 | 256 | Conv4_2 | 3×3 | 512 | ||
| Conv3_2 | 3×3 | 256 | Conv4_1 | 3×3 | 512 | ||
| Conv3_3 | 3×3 | 256 | Conv3_3 | 3×3 | 256 | ||
| Conv4_1 | 3×3 | 512 | Conv3_2 | 3×3 | 256 | ||
| Conv4_2 | 3×3 | 512 | Conv3_1 | 3×3 | 256 | ||
| Conv4_3 | 3×3 | 512 | Conv2_2 | 3×3 | 128 | ||
| Conv5_1 | 3×3 | 512 | Conv2_1 | 3×3 | 128 | ||
| Conv5_2 | 3×3 | 512 | Conv1_2 | 3×3 | 64 | ||
| Conv5_3 | 3×3 | 512 | Conv1_1 | 3×3 | 64 |
Fig 4Pooling and upsampling.
Fig 5Feature map visualization of the convolutional layer.
Fig 6Performance comparison on test images.
(a) original images; (b) ground truth; (c) output by our method; (d) output by FCN; and (e) output by U-Net.
Results of SegNet, FCN and U-Net.
| Approach | MPA | MIoU | FWIoU | Speed |
|---|---|---|---|---|
| 0.927 | 0.618 | 0.844 | 0.604 s | |
| 0.895 | 0.538 | 0.759 | 0.148 s | |
| 0.708 | 0.530 | 0.831 | 0.331 s |
Comparison of the SegNet, FCN and U-Net approaches.
| Approach | GT/Predicted Class | Rice | Background | Weed |
|---|---|---|---|---|
| Rice | 0.015 | 0.049 | ||
| Background | 0.061 | 0.032 | ||
| Weed | 0.042 | 0.019 | ||
| Rice | 0.022 | 0.056 | ||
| Background | 0.120 | 0.046 | ||
| Weed | 0.053 | 0.018 | ||
| Rice | 0.182 | 0.356 | ||
| Background | 0.014 | 0.010 | ||
| Weed | 0.173 | 0.143 |