| Literature DB >> 31683734 |
Muhammad Hammad Saleem1, Johan Potgieter2, Khalid Mahmood Arif3.
Abstract
Plant diseases affect the growth of their respective species, therefore their early identification is very important. Many Machine Learning (ML) models have been employed for the detection and classification of plant diseases but, after the advancements in a subset of ML, that is, Deep Learning (DL), this area of research appears to have great potential in terms of increased accuracy. Many developed/modified DL architectures are implemented along with several visualization techniques to detect and classify the symptoms of plant diseases. Moreover, several performance metrics are used for the evaluation of these architectures/techniques. This review provides a comprehensive explanation of DL models used to visualize various plant diseases. In addition, some research gaps are identified from which to obtain greater transparency for detecting diseases in plants, even before their symptoms appear clearly.Entities:
Keywords: convolutional neural networks (CNN); deep learning; plant disease
Year: 2019 PMID: 31683734 PMCID: PMC6918394 DOI: 10.3390/plants8110468
Source DB: PubMed Journal: Plants (Basel) ISSN: 2223-7747
Figure 1Summary of the evolution of deep learning from 1943–2006.
Figure 2Flow diagram of DL implementation: First, the dataset is collected [25] then split into two parts, normally into 80% of training and 20% of validation set. After that, DL models are trained from scratch or by using transfer learning technique, and their training/validation plots are obtained to indicate the significance of the models. Then, performance metrics are used for the classification of images (type of particular plant disease), and finally, visualization techniques/mappings [55] are used to detect/localize/classify the images.
Figure 3Summary of the evolution of various deep learning models from 2012 until now.
Comparison of state-of-the-art deep learning models.
| Deep Learning Models | Parameters | Key Features and Pros/Cons |
|---|---|---|
| LeNet | 60k | First CNN model. Few parameters as compared to other CNNmodels. Limited capability of computation |
| AlexNet | 60M | Known as the first modern CNN. Best image recognition performance at its time. Used ReLU to achieve better performance. Dropout technique was used to avoid overfitting |
| OverFeat | 145M | First model used for detection, localization, and classification of objects through a single CNN. Large number of parameters as compared to AlexNet |
| ZFNet | 42.6M | Reduced weights (as compared to AlexNet) by considering 7 × 7 kernels and improved accuracy |
| VGG | 133M–144M | 3 × 3 receptive fields were considered to include more number of non-linearity functions which made decision function discriminative. Computationally expensive model due to large number of parameters |
| GoogLeNet | 7M | Fewer number of parameters as compared to AlexNet model. Better accuracy at its time |
| ResNet | 25.5M | Vanishing gradient problem was addressed. Better accuracy than VGG and GoogLeNet models |
| DenseNet | 7.1M | Dense connections between the layers. Reduced number of parameters with better accuracy |
| SqueezeNet | 1.25M | Similar accuracy as AlexNet with 50 times lesser parameters. Considered 1 × 1 filters instead of 3 × 3 filters. Input channels were decreased. Large activation maps of convolution layers |
| Xception | 22.8M | A depth-wise separable convolution approach. Performed better than VGG, ResNet, and Inception-v3 models |
| MobileNet | 4.2M | Considered the depth-wise separable convolution concept. Reduced parameters significantly. Achieved accuracy near to VGG and GoogLeNet |
| Modified/Reduced MobileNet | 0.5/0.54M | Lesser number of parameters as compared to MobileNet. Similar accuracy as compared to MobileNet |
| VGG-Inception | 132M | A cascaded version of VGG and inception module. The number of parameters were reduced by substituting 5 × 5 convolution layers with two 3 × 3 layers. Testing accuracy was increased as compared to many well-known DL models like AlexNet, GoogLeNet, Inception-v3, ResNet, and VGG-16. |
Visualization mapping/techniques used in several approaches.
| Visualization Techniques/Mappings | References |
|---|---|
| Visualization of features having filter from first to final layer | [ |
| Visualize activations in first convolutional layer | [ |
| Saliency map visualization | [ |
| Classification and localization of diseases by bounding boxes | [ |
| Heat maps were used to identify the spots of the disease | [ |
| Feature map for the diseased rice plant | [ |
| Symptoms visualization method | [ |
| Feature and spatial core maps | [ |
| Color space into HSV and K-means clustering | [ |
| Feature map for spotting the diseases | [ |
| Image segmentation method | [ |
| Reconstruction of images on discriminant regions, segmentation of images by binary threshold theorem, and heat map construction | [ |
| Saliency map visualization | [ |
| Saliency map, 2D and 3D contour, mesh graph image | [ |
| Activation visualization | [ |
| Segmentation map and edge map | [ |
Figure 4Feature maps after the application of convolution to an image: (a) real image, (b) first convolutional layer filter, (c) rectified output from first layer, (d) second convolutional layer filter, (e) output from second layer, (f) output of third layer, (g) output of fourth layer, (h) output of fifth layer [27].
Figure 5Tomato plant disease detection by heat map: on left hand side (a) tomato early blight, (b) tomato septoria leaf spot, (c) tomato late blight and (d) tomato leaf mold) and saliency map; on right hand side (a) tomato healthy, (b) tomato late blight, (c) tomato early blight, (d) tomato septoria leaf spot, (e) tomato early blight, (f) tomato leaf mold) [55].
Figure 6Detection of maize disease (indicated by red circles) by heat map [70].
Figure 7Bounding box indicates the type of diseases along with the probability of their occurrence [68]. A bounding box technique was used in Figure 7 in which (a) represents the one type of disease along with its rate of occurrence, (b) indicates three types of plant disease (miner, temperature, and gray mold) in a single image, (c,d) shows one class of disease but contains different patterns on the front and back side of the image, (e,f) displays different patterns of gray mold in the starting and end stages [68].
Figure 8(a) Teacher/student architecture approach; (b) segmentation using a binary threshold algorithm [67].
Figure 9Comparison of Teacher/student approach visualization map with the previous approaches [67].
Figure 10Activation visualization for detection of apple plant disease to show the significance of a VGG-Inception model (the plant disease is indicated by the red circle) [85].
Figure 11Segmentation and edge map for olive leaf disease detection [65].
Figure 12Deep learning models used in the particular number of research papers.
Comparison of several DL approaches in terms of various performance metrics.
| DL Architectures/Algorithms | Datasets | Selected Plant/s | Performance Metrics (and Their Results) | Refs |
|---|---|---|---|---|
| CNN | PlantVillage | Maize | CA (92.85%) | [ |
| AlexNet, GoogLeNet, ResNet | PlantVillage | Tomato | CA by ResNet which gave the best value (97.28%) | [ |
| LeNet | PlantVillage | Banana | CA (98.61%), F1 (98.64%) | [ |
| AlexNet, ALexNetOWTBn, GoogLeNet, Overfeat, VGG | PlantVillage and in-field images | Apple, blueberry, banana, cabbage, cassava, cantaloupe, celery, cherry, cucumber, corn, eggplant, gourd, grape, orange, onion | Success rate of VGG (99.53%) which is the best among all | [ |
| AlexNet, VGG16, VGG 19, SqueezeNet, GoogLeNet, Inceptionv3, InceptionResNetv2, ResNet50, Resnet101 | Real field dataset | Apricot, Walnut, Peach, Cherry | F1(97.14), Accuracy (97.86 ± 1.56) of ResNet | [ |
| Inceptionv3 | Experimental field dataset | Cassava | CA (93%) | [ |
| CNN | Images taken from the research center | Cucumber | CA (82.3%) | [ |
| Super-Resolution Convolutional Neural Network (SCRNN) | PlantVillage | Tomato | Accuracy (~90%) | [ |
| CaffeNet | Downloaded from the internet | Pear, cherry, peach, apple, grapevine | Precision (96.3%) | [ |
| AlexNet and GoogLeNet | PlantVillage | Apple, blueberry, bell pepper, cherry, corn, peach, grape, raspberry, potato, squash, soybean, strawberry, tomato | CA (99.35%) of GoogLeNet | [ |
| AlexNet, GoogLeNet, VGG- 16, ResNet-50,101, ResNetXt-101, Faster RCNN, SSD, R-FCN, ZFNet | Image taken in real fields | Tomato | Precision (85.98%) of ResNet-50 with Region based Fully Convolutional Network(R-FCN) | [ |
| CNN | Bisque platform of Cy Verse | Maize | Accuracy (96.7%) | [ |
| DCNN | Images were taken in real field | Rice | Accuracy (95.48%) | [ |
| AlexNet, GoogLeNet | PlantVillage | Tomato | Accuracy (0.9918 ± 0.169) of GoogLeNet | [ |
| VGG-FCN-VD16 and VGG-FCN-S | Wheat Disease Database 2017 | Wheat | Accuracy (97.95%) of VGG-FCN-VD16 | [ |
| VGG-A, CNN | Images were taken in real field | Radish | Accuracy (93.3%) | [ |
| AlexNet | Images were taken in real field | Soybean | CA (94.13%) | [ |
| AlexNet and SqueezeNet v1.1 | PlantVillage | Tomato | CA (95.65%) of AlexNet | [ |
| DCNN, Random forest, Support Vector Machine and AlexNet | PlantVillage dataset, Forestry Image dataset and agricultural field in China | Cucumber | CA (93.4%) of DCNN | [ |
| Teacher/student architecture | PlantVillage | Apple, bell pepper, blueberry, cherry, corn, orange, grape, potato, raspberry, peach, soybean, strawberry, tomato, squash | Training accuracy and loss (~99%,~0–0.5%), validation accuracy and loss (~95%, ~10%) | [ |
| Improved GoogLeNet, Cifar-10 | PlantVillage and various websites | Maize | Top-1 accuracy (98.9%) of improved GoogLeNet | [ |
| MobileNet, Modified MobileNet, Reduced MobileNet | PlantVillage dataset | 24 types of plant | CA (98.34%) of reduced MobileNet | [ |
| VGG-16, ResNet-50,101,152, Inception-V4 and DenseNets-121 | PlantVillage | Apple, bell pepper, blueberry, cherry, corn, orange, grape, potato, raspberry, peach, soybean, strawberry, tomato, squash | Testing accuracy (99.75%) of DenseNets | [ |
| User defined CNN, SVM, AlexNet, GoogLeNet, ResNet-20 and VGG-16 | Images were taken in real field | Apple | CA (97.62%) of proposed CNN | [ |
| AlexNet and VGG-16 | PlantVillage | Tomato | CA (AlexNet) | [ |
| LeafNet, SVM, MLP | Images were taken in real field | Tea leaf | CA (90.16%) of LeafNet | [ |
| 2D-CNN-BidGRU | Real wheat field | wheat | F1 (0.75) and accuracy (0.743) | [ |
| OR-AC-GAN | Real environment | Tomato | Accuracy (96.25%) | [ |
| 3D CNN | Real environment | Soybean | CA (95.73%), F1-score (0.87) | [ |
| DCNN | Real environment | Wheat | Accuracy (85%) | [ |
| ResNet-50 | Real environment | Wheat | Balanced Accuracy (87%) | [ |
| GPDCNN | Real environment | Cucumber | CA (94.65%) | [ |
| VGG-16, AlexNet | PlantVillage, CASC-IFW | Apple, banana | CA (98.6%) | [ |
| LeNet | Real environment | Grapes | CA (95.8%) | [ |
| PlantDiseaseNet | Real environment | Apple, bell-pepper, cherry, grapes, onion, peach, potato, plum, strawberry, sugar-beets, tomato, wheat | CA (93.67%) | [ |
| LeNet | PlantVillage | Soybean | CA (99.32%) | [ |
| VGG-Inception | Real environment | Apple | Mean average accuracy (78.8%) | [ |
| Resnet-50, Inception-V2, MobileNet-V1 | Real environment | Banana | Mean average accuracy (99%) of ResNet-50 | [ |
| Modified LeNet | PlantVillage | Olives | True positive rate (98.6 ± 1.47%) | [ |
Figure 13Sample images of OR-AC-GAN (a hyperspectral imaging model) [112].
Figure 14Hyperspectral images by UAV: (a) RGB color plots, (b) Random-Forest classifier, and (c) proposed multiple Inception-ResNet model [114].