| Literature DB >> 35574075 |
Bishwa B Sapkota1, Chengsong Hu1,2, Muthukumar V Bagavathiannan1.
Abstract
Convolutional neural networks (CNNs) have revolutionized the weed detection process with tremendous improvements in precision and accuracy. However, training these models is time-consuming and computationally demanding; thus, training weed detection models for every crop-weed environment may not be feasible. It is imperative to evaluate how a CNN-based weed detection model trained for a specific crop may perform in other crops. In this study, a CNN model was trained to detect morningglories and grasses in cotton. Assessments were made to gauge the potential of the very model in detecting the same weed species in soybean and corn under two levels of detection complexity (levels 1 and 2). Two popular object detection frameworks, YOLOv4 and Faster R-CNN, were trained to detect weeds under two schemes: Detect_Weed (detecting at weed/crop level) and Detect_Species (detecting at weed species level). In addition, the main cotton dataset was supplemented with different amounts of non-cotton crop images to see if cross-crop applicability can be improved. Both frameworks achieved reasonably high accuracy levels for the cotton test datasets under both schemes (Average Precision-AP: 0.83-0.88 and Mean Average Precision-mAP: 0.65-0.79). The same models performed differently over other crops under both frameworks (AP: 0.33-0.83 and mAP: 0.40-0.85). In particular, relatively higher accuracies were observed for soybean than for corn, and also for complexity level 1 than for level 2. Significant improvements in cross-crop applicability were further observed when additional corn and soybean images were added to the model training. These findings provide valuable insights into improving global applicability of weed detection models.Entities:
Keywords: CNNs; deep learning; digital technologies; precision agriculture; precision weed control; site-specific weed management
Year: 2022 PMID: 35574075 PMCID: PMC9096552 DOI: 10.3389/fpls.2022.837726
Source DB: PubMed Journal: Front Plant Sci ISSN: 1664-462X Impact factor: 5.753
Figure 1(A) Study area (Texas A&M AgriLife Research Farm, Burleson County, TX, United States) and field setup for the 2 experimental years; (B) a multi-copter drone (Hylio Inc., Houston, TX, United States) attached with Fujifilm GFX100 (100 MP) camera; and (C) image datasets (top and bottom rows) collected under two different environmental conditions for cotton, soybean, and corn.
Various datasets used in the study.
| Image dataset name | Acquisition date | Crop/growth stage | Weed composition/growth stage | Weed density (plants m−2) | Image acquisition conditions | Train/Val/Test [images, annotations] | Annotation composition | |
|---|---|---|---|---|---|---|---|---|
| 1 | Cotton 1 (Test data referred to as Cot1) | May 06, 2020 | Cotton: 4–5 leaves | MG: cotyledon-4 leaves | 18 | Sunny | Train: [460, 8,580] | [19.3, 79.5, 1.2] |
| 2 | Cotton 2 (referred to as Cot2) | June 13, 2021 | Cotton: 2–4 leaves | MG: cotyledon-6 leaves | 21 | Partially cloudy | Test: [95, 600] | [36, 63.8, 0.2] |
| 3 | Soybean 1 (Test data referred to as Soy1) | May 06, 2020 | Soybean: 6–7 leaves | MG: cotyledon-4 leaves | 17 | Sunny | Train: [115, 990] | [46.4, 53.48, 0.07] |
| 4 | Soybean 2 (referred to as Soy2) | May 14, 2021 | Soybean: 1–3 leaves | MG: cotyledon-6 leaves | 21 | Cloudy | Test: [97, 547] | [63.07, 35.4, 1.53] |
| 5 | Corn 1 (Test data referred to as Corn1) | May 07, 2021 | Corn: 2–3 leaves | MG: cotyledon-3 leaves | 18 | Sunny | Train: [115, 1,010] | [81.16, 16.75, 2.1] |
| 6 | Corn 2 (referred to as Corn2) | May 14, 2021 | Corn: 3–4 leaves | MG: cotyledon-6 leaves | 23 | Cloudy | Test: [95, 559] | [80.5, 17.5, 2] |
Train, training; Val, validation; MG, morningglories; TM, texas millet; and JG, johnsongrass.
The annotations statistics shown within the brackets are given in %.
Figure 2Schematic showing the workflow used in the study. The study began with data collection using an UAV and the collected data were distributed for training and test datasets. Data management was followed by model training under two detection schemes: Detect_Weed (detecting at weed/crop level) and Detect_Species (detecting at weed species level). After the models were trained, they were evaluated on the test datasets (Other was excluded during the calculation of accuracy metrics). Average Precision (AP) and Mean Average Precision (mAP) was used as the metrics for performance evaluation.
Various training datasets evaluated in the study for training YOLOv4 and Faster R-CNN and annotations record for each training dataset.
| Training dataset | Non-cotton images (%) | Annotations | |||
|---|---|---|---|---|---|
| MG (%) | Grass (%) | Other (%) | Total | ||
| Train100 | 0 | 19.3 | 79.5 | 1.20 | 8,580 |
| Train105 | 5 | 20.0 | 78.8 | 1.25 | 8,775 |
| Train110 | 10 | 20.7 | 78.0 | 1.23 | 8,915 |
| Train115 | 15 | 21.7 | 77.1 | 1.21 | 9,072 |
| Train120 | 20 | 22.9 | 75.9 | 1.19 | 9,234 |
| Train125 | 25 | 23.0 | 75.7 | 1.33 | 9,480 |
| Train130 | 30 | 23.9 | 74.7 | 1.32 | 9,689 |
| Train135 | 35 | 24.3 | 74.4 | 1.34 | 9,827 |
| Train140 | 40 | 25.0 | 73.7 | 1.32 | 9,970 |
| Train145 | 45 | 25.5 | 73.2 | 1.31 | 10,113 |
| Train150 | 50 | 25.7 | 73.0 | 1.29 | 10,198 |
MG-Morningglories; Grass-Grass weeds; and Other-Weeds other than MG and Grass.
The numerical figures in this column indicate the percentage of images added to Train100 (i.e., 460 images).
Train100 had a total of 460 cotton images and 0 non-cotton images.
“Train100” represents the dataset with cotton images only, i.e., no non-cotton images. The last two digits of training dataset names represent the percentage of non-cotton images added to Train100 randomly for building the respective training dataset. The percentage was with respect to Train100.
Accuracy obtained for various test datasets with YOLOv4 and Faster R-CNN under Detect_Weed and Detect_Species using the main cotton model.
| Detect_Weed | Detect_Species | |||||||
|---|---|---|---|---|---|---|---|---|
| YOLOv4 | Faster R-CNN | YOLOv4 | Faster R-CNN | |||||
| AP | AP | AP (MG) | AP (Grass) | mAP | AP (MG) | AP (Grass) | mAP | |
| Cot1 | 0.88 | 0.87 | 0.88 | 0.83 | 0.85 | 0.86 | 0.79 | 0.83 |
| Cot2 | 0.79 | 0.74 | 0.71 | 0.79 | 0.75 | 0.60 | 0.70 | 0.65 |
| Soy1 | 0.83 | 0.76 | 0.83 | 0.75 | 0.79 | 0.72 | 0.70 | 0.71 |
| Soy2 | 0.35 | 0.60 | 0.63 | 0.64 | 0.64 | 0.72 | 0.49 | 0.61 |
| Corn1 | 0.72 | 0.62 | 0.88 | 0.15 | 0.52 | 0.78 | 0.15 | 0.47 |
| Corn2 | 0.33 | 0.39 | 0.65 | 0.15 | 0.40 | 0.54 | 0.03 | 0.29 |
MG-Morningglories; Grass-Grass weeds; AP, average precision; and mAP, mean average precision. AP and mAP values were computed to assess the performance of the main cotton model over the test datasets. mAP was calculated by averaging AP for MG and Grass. AP was calculated as a function of precision and recall values obtained when Intersection Over Union (IoU) was set to 0.5.
Figure 3Weed detection using bounding boxes by the main cotton models under “Detect_Weed” scheme for various test datasets used in the study. YOLOv4 and Faster R-convolutional neural network (CNN) were trained with the Train100 dataset (i.e., dataset containing cotton images only) to develop the main cotton models. Under this scheme, MG, Grass, and Other were combined into “Weed” category while training the model.
Figure 4Bounding boxes generated for MG and Grass by the main cotton models under “Detect_Species” scheme for various test datasets used in the study. YOLOv4 and Faster R-CNN were trained with the Train100 dataset (i.e., dataset containing cotton images only) to develop the main cotton models. Under this scheme, MG, Grass, and Other were trained as separate categories.
Figure 5Average Precision and mAP achieved for different complexity level datasets with main cotton models. Complexity level 1 datasets include Soy1 and Corn1, whereas level 2 include Soy2 and Corn2. The main cotton models were derived by training the detection frameworks (YOLOv4 and Faster R-CNN) with Train100 (i.e., dataset containing cotton images only). The AP/mAP for datasets under each complexity level were averaged to derive average AP and mAP.
Figure 6Line plots showing AP and mAP achieved with various training datasets for each test dataset used in the study for both frameworks and detection schemes. Various training datasets were created by adding Soybean 1 and Corn 1 training images to the original dataset, i.e., Train100. These non-cotton crop images were added 5% at a time until they amounted to 50% of Train100. The last two digits in the training dataset name denote the % of images added to Train100.
The maximum rate of increment in accuracy for various test datasets with the addition of non-cotton images.
| Detect_Weed (AP%) | Detect_Species (mAP%) | |||
|---|---|---|---|---|
| YOLOv4 | Faster R-CNN | YOLOv4 | Faster R-CNN | |
| Cot1 | 2.27 | 2.29 | 5.89 | 2.42 |
| Cot2 | 7.60 | 2.70 | 2.00 | 7.70 |
| Soy1 | 3.61 | 5.26 | 6.32 | 7.74 |
| Soy2 | 122.8 | 16.00 | 11.90 | 8.27 |
| Corn1 | 31.9 | 53.22 | 12.62 | 34.40 |
| Corn2 | 127.27 | 69.23 | 28.75 | 58.62 |
AP, average precision; mAP, mean average precision. The rate was determined by subtracting the accuracy obtained with Train100 (no non-cotton images) from the highest accuracy obtained among all training datasets for the respective test dataset.
Figure 7Line plots showing AP and mAP achieved for each complexity level with YOLOv4 and Faster R-CNN. Complexity level 1 datasets include Soy1 and Corn1, whereas level 2 include Soy2 and Corn2. AP and mAP for Cot1 dataset were also included in the averaging process of each complexity level to understand how well the models perform with both cotton and non-cotton datasets.