| Literature DB >> 31396250 |
Justine Boulent1,2,3, Samuel Foucher2, Jérôme Théau1,3, Pierre-Luc St-Charles2.
Abstract
Deep learning techniques, and in particular Convolutional Neural Networks (CNNs), have led to significant progress in image processing. Since 2016, many applications for the automatic identification of crop diseases have been developed. These applications could serve as a basis for the development of expertise assistance or automatic screening tools. Such tools could contribute to more sustainable agricultural practices and greater food production security. To assess the potential of these networks for such applications, we survey 19 studies that relied on CNNs to automatically identify crop diseases. We describe their profiles, their main implementation aspects and their performance. Our survey allows us to identify the major issues and shortcomings of works in this research area. We also provide guidelines to improve the use of CNNs in operational contexts as well as some directions for future research.Entities:
Keywords: Review (article); convolutional neural networks; deep learning; plant diseases detection; precision agriculture
Year: 2019 PMID: 31396250 PMCID: PMC6664047 DOI: 10.3389/fpls.2019.00941
Source DB: PubMed Journal: Front Plant Sci ISSN: 1664-462X Impact factor: 5.753
Figure 1Expected output examples of (A) the classification, (B) the object detection, and (C) the segmentation of images containing esca disease symptoms.
Figure 2Typology of image complexity found in the datasets. Esca grape disease on (A) an image captured under controlled condition (from the PlantVillage dataset), (B) an image captured under uncontrolled condition and with a focus on a particular organ, and (C) an image captured under uncontrolled conditions and without focus on a particular organ.
Figure 3Selecting an analysis scale: from a scale close up on the main symptoms (A) to a scale providing more contextual features (D). Example of a vine branch affected by flavescence dorée.
Figure 4Comparison of the training time (h) and accuracy values (%) obtained on a validation set according to different architectures and training strategies. Adapted from Brahimi et al. (2018).
Figure 5Comparison of the accuracies obtained using different architectures. For each study, the lowest and highest performance achieved on the validation set (or test set when available) is reported.
Figure 6Comparison of accuracy values obtained by CNNs and by other image processing methods. Only the best CNN architecture and comparative approach for each study were reported. If given in the study, the precision on a test set was reported.
Figure 7Visualization examples found in the corpus (A) Activations in the first convolution layer visualization (Mohanty et al., 2016). (B) T-distributed Stochastic Neighbor Embedding on the final fully connected layer (Zhang K. et al., 2018).
| Atabay, | Tomato | PlantVillage subset | 10 | 19,742 | 373–5,357 | |
| Barbedo, | 12 crop plants | Barbedo, | 56 | 1,383 | 5–77 | |
| Brahimi et al., | Tomato | PlantVillage subset | 9 | 14,828 | 325–4,032 | |
| Brahimi et al., | 14 crop species | PlantVillage | 39 | 54,323 | 152–5,507 | |
| Cruz et al., | Olive | Own dataset (controlled) | 3 | 299 | 99–100 | |
| DeChant et al., | Maize | Own dataset (field) | 2 | 1,796 | 768–1,028 | |
| Ferentinos, | 25 crop species | PlantVillage dataset | 58 | 87,848 | 43–6,235 | |
| Fuentes et al., | Tomato | Own dataset (field) | 10 | 5,000 | 338–18,899 | |
| Fuentes et al., | Tomato | Own dataset (field) | 12 | 8,927 | 338–18,899 | |
| Liu B. et al., | Apple | Own dataset (controlled and field) | 4 | 1,053 | 2,366–4,147 | |
| Mohanty et al., | 14 crop species | PlantVillage | 38 | 54,306 | 152–5,507 | |
| Oppenheim and Shani, | Potato | Own dataset (controlled) | 5 | 400 | 265–738 | |
| Picon et al., | Wheat | Johannes et al., | 4 | 8,178 | 1,116–3,338 | |
| Ramcharan et al., | Cassava | Own dataset (field) | 6 | 2,756 | 309–415 | |
| Sladojevic et al., | Apple, Pear, Cherry, Peach, Grapevine | Own dataset (internet) | NA | 15 | 4,483 | 108–1,235 |
| Too et al., | 14 crop plants | PlantVillage | 38 | 54,306 | 152–5,507 | |
| Wang et al., | Apple | PlantVillage subset | 4 | 2,086 | 145–1,644 | |
| Zhang S. et al., | A/ Tomato B/ Cucumber | A/ PlantVillage subset B/ Own dataset (in field) | A/ * B/ ** | A/ 8 B/ 5 | A/ 15,817 B/ 500 | A/ 366 - 5350 B/ 100 |
| Zhang K. et al., | Tomato | PlantVillage subset | 9 | 5,550 | 405–814 |
NA, Not Applicable.
Complexity level allowing identification under controlled conditions.
Complexity level allowing identification under uncontrolled conditions.
Complexity level allowing the development of automatic screening tools.
| Atabay, | C | VGG16, 19, custom architecture | FS–TL | 97.53 | ** |
| Barbedo, | C | GoogleNet | TL | 87 | * |
| Brahimi et al., | C | AlexNet, GoogleNet | FS–TL | 99.18 | * |
| Brahimi et al., | C | AlexNet, DenseNet169, Inception v3, ResNet34, SqueezeNet1-1.1, VGG13 | FS -TL | 99.76 | * |
| Cruz et al., | C | LeNet | TL | 98.60 | * |
| DeChant et al., | D | Custom three stages architecture | FS | 96.70 | ** |
| Ferentinos, | C | AlexNet, AlexNetOWTBn, GoogleNet, Overfeat, VGG | Unspecified | 99.53 | * |
| Fuentes et al., | D | AlexNet, ZFNet, GoogleNet, VGG16, ResNet50, 101, ResNetXt-101 | TL | 85.98 | ** |
| Fuentes et al., | D | Custom architecture with Refinement Filter Bank | TL | 96.25 | ** |
| Liu B. et al., | C | AlexNet, GoogleNet, ResNet 20, VGG 16 and custom architecture | FS -TL | 97.62 | * |
| Mohanty et al., | C | AlexNet, GoogleNet | FS–TL | 31 | *** |
| Oppenheim and Shani, | C | VGG | Unspecified | 96 | * |
| Picon et al., | C | Custom ResNet50, Resnet50 | TL | 97 | *** |
| Ramcharan et al., | C | Inception V3 | TL | 93 | ** |
| Sladojevic et al., | C | CaffeNet | TL | 96.3 | * |
| Too et al., | C | Inception V4, VGG 16, ResNet 50, 101 and 152, DenseNet 121 | TL | 99.75 | ** |
| Wang et al., | C | VGG16, 19, Inception-V3, ResNet50 | TL | 90.40 | * |
| Zhang S. et al., | C | Custom Three Channels CNN, DNN, LeNet-5, GoogleNet | FS | A/ 87.15 B/ 91.16 | A/ * B/ * |
| Zhang K. et al., | C | AlexNet, GoogleNet, ResNet | TL | 97.28 | * |
Classification (C)—Detection (D).
From Scratch (FS)—Transfer Learning (TL).
If available, the accuracy of the explicitly different test set is privileged.
SSAbsence of three explicit subsets; SSThree explicit subsets; SSSTest set explicitly different from the training set.