| Literature DB >> 35893646 |
Hamish A Craze1, Nelishia Pillay2, Fourie Joubert1, Dave K Berger3.
Abstract
Maize yields worldwide are limited by foliar diseases that could be fungal, oomycete, bacterial, or viral in origin. Correct disease identification is critical for farmers to apply the correct control measures, such as fungicide sprays. Deep learning has the potential for automated disease classification from images of leaf symptoms. We aimed to develop a classifier to identify gray leaf spot (GLS) disease of maize in field images where mixed diseases were present (18,656 images after augmentation). In this study, we compare deep learning models trained on mixed disease field images with and without background subtraction. Performance was compared with models trained on PlantVillage images with single diseases and uniform backgrounds. First, we developed a modified VGG16 network referred to as "GLS_net" to perform binary classification of GLS, which achieved a 73.4% accuracy. Second, we used MaskRCNN to dynamically segment leaves from backgrounds in combination with GLS_net to identify GLS, resulting in a 72.6% accuracy. Models trained on PlantVillage images were 94.1% accurate at GLS classification with the PlantVillage testing set but performed poorly with the field image dataset (55.1% accuracy). In contrast, the GLS_net model was 78% accurate on the PlantVillage testing set. We conclude that deep learning models trained with realistic mixed disease field data obtain superior degrees of generalizability and external validity when compared to models trained using idealized datasets.Entities:
Keywords: Cercospora; crop disease; deep learning; field conditions; gray leaf spot; maize; plant pathology
Year: 2022 PMID: 35893646 PMCID: PMC9330607 DOI: 10.3390/plants11151942
Source DB: PubMed Journal: Plants (Basel) ISSN: 2223-7747
Deep learning applications for plant disease classification.
| Plant Species | Disease | Dataset | Size | Architecture | Highest Accuracy | References |
|---|---|---|---|---|---|---|
| Apple | Black Rot | PlantVillage | 2086 | VGG-16, VGG-19, Inception-v3, ResNet50 | 90.4% | [ |
| Maize | Northern Corn Leaf Blight | Manual * | 1796 | CNN | 96.7% | [ |
| Maize | Northern Corn Leaf Blight | Manual | 3000 | MaskRCNN | 96% (AP) * | [ |
| Maize | Common Rust | PlantVillage | 1800 | VGG-16 | 89% | [ |
| Maize | Southern Leaf Blight, Brown Spot, Curvularia Leaf Spot, Rust, Dwarf Mosaic, Gray Leaf Spot, Round Spot, Northern Leaf Blight | PlantVillage and Various | 500 | GoogLeNet | 98.8% | [ |
| Maize | Common Rust, Gray Leaf Spot, Northern Corn Leaf Blight, Healthy | PlantVillage | 3852 | Modified LeNet | 97.89% | [ |
| Maize | Rust, Nothern Corn Leaf Blight, Healthy | Manual in Tandem with PlantVillage | 4382 | Custom DCNN | 88.46% | [ |
| Pear, cherry, peach, apple, grapevine | 7 diseases (fungal, oomycete, bacterial, mites | Various | 30,880 | CaffeNet | 96.3% | [ |
| Potato | Potato Blight | PlantVillage | 300 | SVM | 95% | [ |
| Soybean | 4 diseases (fungal, bacterial), 3 abiotic stresses | Manual, But Highly Controlled | 6000 | DCNN | 94.13% | [ |
| Tomato | One bacterial, two viruses, five fungal diseases, spider mites | PlantVillage | 14,828 | AlexNet, GoogLeNet and others | 99.18% | [ |
| Tomato | 5 diseases (fungal, oomycete, bacterial), 2 insects, 2 abiotic factors | Manual | 5000 | Faster R-CNN, R-FCN, SSD | 85.98% | [ |
| Wheat | Powdery Mildew, Smut, Black Chaff, Stripe Rust, Leaf Blotch, Leaf Rust, Healthy Wheat | WDD2017 | 9230 | VGG-FCN-VD16, VGG-FCN-S | 95.12% | [ |
| 14 crops (dicots, trees monocots) | 38 Diseases (fungal, oomycete, bacterial, viral) | PlantVillage | 54,306 | AlexNet, GoogLeNet | 99.35% | [ |
* Manual = image dataset developed by authors. * AP = Average precision. MaskRCNN networks are not assessed using accuracy.
Figure 1Images of maize leaves obtained from PlantVillage. (a) Image labelled as GLS positive. (b) Image labelled as CR positive, note the presence of Phaeosphaeria Leaf Spot (PLS). (c) Image labelled as NCLB positive, note the presence of CR. (d) Image labelled as ‘Healthy’.
Figure 2Example images of GLS symptoms on maize leaves in the In-field (IF) dataset. (a) Example of GLS lesions as visualized from under the leaf. (b) Example of GLS coalescing into larger, differently shaped lesions. (c) Example of a GLS lesion (red) occurring inside an NCLB lesion (blue).
Breakdown of disease classes found in the In-field dataset.
| Disease | Total |
|---|---|
| Gray Leaf Spot (GLS) | 1084 |
| Northern Corn Leaf Blight (NCLB) | 554 |
| Phaeosphaeria Leaf Spot (PLS) * | 493 |
| Common Rust (CR) | 300 |
| Southern Rust (SR) | 39 |
| No Foliar Symptoms | 285 |
| Other | 324 |
| Unidentified | 309 |
| Total Images | 2332 |
| Total Disease observations | 3388 |
* Also known as White Spot Disease.
Extent of disease co-occurrence in the In-field (IF) dataset.
| Number of Classes per Image | Total |
|---|---|
| 1 | 1415 |
| 2 | 48 |
| 3 | 691 |
| 4 | 31 |
| 5 | 128 |
| 6 | 13 |
| 7 | 19 |
| 8 | 0 |
| AVG number of classes per image | 1.45 |
| STD of number of classes per image | 0.63 |
Figure 3Example of image segmentation to define the leaf area for the “In-field_leaf” (IFL) dataset. (a) Original leaf image from the In-field (IF) dataset. (b) Leaf area from image (a) highlighted manually and shown by brown overlay using the tool available at https://segments.ai (accessed on 21 July 2022).
Figure 4(a) Image of a maize leaf and (b) the same leaf after leafRCNN leaf area prediction and background removal.
Breakdown of classes found in the PV dataset.
| Disease | Total |
|---|---|
| Gray Leaf Spot | 513 |
| Northern Corn Leaf Blight | 1192 |
| Common Rust | 985 |
| Healthy | 1162 |
| Total | 3852 |
Performance of GLS_net upon the IF testing set.
| Name | Precision | Recall | F1-Score | Support |
|---|---|---|---|---|
| GLS | 86.3 | 50.0 | 63.3 | 2136 |
| notGLS | 68.8 | 93.3 | 79.2 | 2528 |
| Macro Avg | 77.5 | 71.6 | 71.2 | 4664 |
| Weighted Avg | 76.8 | 73.4 | 71.9 | 4664 |
| Accuracy | 73.4 |
Values are provided as percentages (%). Precision = TP/(TP + FP), Recall = TP/(TP + FN), Accuracy = (TP + TN)/(TP + TN + FP + FN). TP = true positive; FP = false positive; TN = true negative; FN = false negative. F1-Score = 2 × (Precision × Recall)/(Precision + Recall). Macro Avg: verage score of metric assuming equal weighting (cannot be calculated from this table, requires underlying data). Weighted Avg: Average weighted score of metric. Metrics are weighted according to class proportion (cannot be calculated from this table, requires underlying data). Support: The total number of images associated with the class.
Performance of GLS_net upon the PV testing set.
| Name | Precision | Recall | F1-Score | Support |
|---|---|---|---|---|
| GLS | 37.5 | 65.7 | 47.7 | 143 |
| notGLS | 93.1 | 80.9 | 86.6 | 820 |
| Macro Avg | 65.3 | 73.3 | 67.1 | 963 |
| Weighted Avg | 84.9 | 78.6 | 80.8 | 963 |
| Accuracy | 78.6 |
Performance of GLS_net_pv upon the PV testing set.
| Name | Precision | Recall | F1-Score | Support |
|---|---|---|---|---|
| GLS | 89.8 | 67.8 | 77.3 | 143 |
| notGLS | 94.6 | 98.7 | 96.6 | 820 |
| Macro Avg | 92.2 | 83.2 | 86.9 | 963 |
| Weighted Avg | 93.9 | 94.1 | 93.7 | 963 |
| Accuracy | 94.1 |
Performance of GLS_net_pv upon the IF testing set.
| Name | Precision | Recall | F1-Score | Support |
|---|---|---|---|---|
| GLS | 61.2 | 5.2 | 9.7 | 2136 |
| notGLS | 54.8 | 97.2 | 70.1 | 2528 |
| Macro Avg | 58.0 | 51.2 | 39.9 | 4664 |
| Weighted Avg | 57.7 | 55.1 | 42.4 | 4664 |
| Accuracy | 55.1 |
leafRCNN performance upon the IFL testing set.
| Metric | IoU Range | Score |
|---|---|---|
| Bbox Precision | 0.50:0.95 | 99.0% |
| Bbox Recall | 0.50:0.95 | 99.0% |
| Segm Precision | 0.50:0.95 | 92.3% |
| Segm Recall | 0.50:0.95 | 94.4% |
Bbox = Bounding Box. MaskRCNN predicts bounding boxes where it believes instances to be contained within. These metrics track how well leafRCNN predicts bounding boxes that overlap with the ground truth (GT). Segm = Segmentation. MaskRCNN predicts masks that should overlay with GT labels. These metrics track how well these predicted masks overlap with GT. Precision = The average precision value obtained between multiple IoU values. Recall = The average recall value obtained between multiple IoU values. IoU = Intersection over Union. Measures the degree of overlap between two 2D objects. 0.50:0.95 indicates that the obtained Precision and Recall values were generated over a range of IoU values between 0.50 and 0.95 using a 0.05 step. Further metrics of leafRCNN performance are given in Tables S1 and S2.
Performance of GLS_net_noBackground upon the IFNB testing set.
| Name | Precision | Recall | F1-Score | Support |
|---|---|---|---|---|
| GLS | 85.0 | 48.8 | 62.0 | 2136 |
| notGLS | 68.2 | 92.7 | 78.6 | 2528 |
| Macro Avg | 76.6 | 70.8 | 70.3 | 4664 |
| Weighted Avg | 75.9 | 72.6 | 71.0 | 4664 |
| Accuracy | 72.6 |
Performance of GLS_net_noBackground upon the PV testing set.
| Name | Precision | Recall | F1-Score | Support |
|---|---|---|---|---|
| GLS | 35.6 | 75.5 | 48.4 | 143 |
| notGLS | 94.7 | 76.2 | 84.5 | 820 |
| Macro Avg | 65.2 | 75.9 | 66.4 | 963 |
| Weighted Avg | 85.9 | 76.1 | 79.1 | 963 |
| Accuracy | 76.1 |
Figure 5Heatmaps from Grad-CAM and Grad-CAM++ software, which are designed to illustrate image regions detected as GLS positive by a CNN, such as GLS_net, are shown here. Panels (a–c) show three GLS positive representative images from the IF dataset. Each panel shows (from left to right) the input image that was scored as GLS positive by GLS_net, the Grad-CAM heatmap, and the Grad-CAM++ heatmap, respectively. Panel (d) contains a colour scale to aid in interpretation, blue indicates no activation, while red indicates high levels of activation.