| Literature DB >> 36124118 |
Zhuoxin Li1, Cong Li1, Linfan Deng1, Yanzhou Fan1, Xianyin Xiao1, Huiying Ma1, Juan Qin1, Liangliang Zhu1.
Abstract
Timely disease detection and pest treatment are key issues in modern agricultural production, especially in large-scale crop agriculture. However, it is very time and effort-consuming to identify plant diseases manually. This paper proposes a deep learning model for agricultural crop disease identification based on AlexNet and Inception-V4. AlexNet and Inception-V4 are combined and modified to achieve an efficient but good performance. Experimental results on the expanded PlantVillage dataset show that the proposed model outperforms the compared methods: AlexNet, VGG11, Zenit, and VGG16, in terms of accuracy and F1 scores. The proposed model obtains the highest accuracy for corn, tomato, grape, and apple: 94.5%, 94.8%, 92.3%, and 96.5%, respectively. Also, the highest F1 scores for corn, tomato, grape, and apple: 0.938, 0.910, 0.945, and 0.924, respectively, are obtained. The results indicate that the proposed method has promising generalization ability in crop disease identification.Entities:
Mesh:
Year: 2022 PMID: 36124118 PMCID: PMC9482484 DOI: 10.1155/2022/5862600
Source DB: PubMed Journal: Comput Intell Neurosci
Summary of related works on plant disease classification.
| Reference | Plant types | Dataset | Data augmentation | Methods | Limitation |
|---|---|---|---|---|---|
| Srdjan et al. [ | 13 kinds of plants | Stanford background dataset | Image transformations used for augmentation: (a)affine transformations; (b)perspective transformations; (c) rotations. | CNN | Training less data |
|
| |||||
| Mohanty et al. [ | 14 crop diseases | PlantVillage | Resize the images to 256 × 256 pixels, and perform both the model optimization and predictions on these downscaled images | AlexNet | When tested on a set of images taken under conditions different from the train images, the accuracy is reduced substantially to just above 31% |
|
| |||||
| Dyrmann et al. [ | 22 crop samples | BBCH12e16 | — | DCNN | Due to the small number of training samples, the recognition accuracy fluctuates greatly |
|
| |||||
| Ferreira et al. [ | Soybean crops diseases | Captured by the UAV | — | ConvNets or CNNs | Dependency on feature extractors |
|
| |||||
| Ghazi et al. [ | 1,000 species of trees, herbs, and ferns | LifeCLEF 2015 | Decrease the chance of overfitting, image transforms such as rotation, translation, reflection, and scaling | GoogleNet, AlexNet, and VGGNet | As an example, increasing the batch size from 20 to 60 increases the training time 3-fold but does not match the performance obtained by increasing the number of iterations by the same amount |
|
| |||||
| Liu et al. [ | 16 kinds of insect pests | Multi-class pest dataset 2018 (MPD2018) | — | CNN | The model did not do a good job of identifying similar pests in different categories methods |
|
| |||||
| Geetharamani and Arun Pandian [ | 13 different of plant leaves | PlantVillage | Image flipping, gamma correction, noise injection, PCA color augmentation, rotation, and scaling transformations | Deep CNN | The model can only identify leaf diseases, but it cannot identify other parts of the plant diseases |
|
| |||||
| Ozguven and Adem [ | Sugar beet leaf disease | Sugar beet leaf images dataset | — | Faster R–CNN | The accuracy of disease detection is low |
|
| |||||
| Chao et al. [ | Apple tree leaf diseases | Laboratory independent planting and cultivation | Image scaling, dataset expansion, and dataset normalization | DCNN | There are few types of data sets, and the specific network architecture of various structures lacks a description |
Figure 1Plant village dataset classification examples.
Dataset of plant leaves.
| Disease type | Label | Total number of pictures (sheets) | Training set (sheets) | Validation set (sheets) |
|---|---|---|---|---|
| Corn | Corn_healthy | 1162 | 930 | 232 |
| Corn_CercosporaGrayspot | 1000 | 800 | 200 | |
| Corn_Commonrust | 1192 | 954 | 238 | |
| Corn_NorthernBlight | 1000 | 800 | 200 | |
|
| ||||
| Tomato | Tomato_healthy | 1591 | 1273 | 318 |
| Tomato_Bacterialspot | 2127 | 1702 | 425 | |
| Tomato_Lateblight | 1908 | 1526 | 382 | |
| Tomato_Septorialspot | 1771 | 1416 | 355 | |
|
| ||||
| Grape | Grape_healthy | 1000 | 800 | 200 |
| Grape_Blackrot | 1180 | 944 | 236 | |
| Grape_IsariopsisSpot | 1076 | 860 | 216 | |
| Grape_Measles | 1383 | 1106 | 277 | |
|
| ||||
| Apple | Apple_healthy | 1645 | 1316 | 329 |
| Apple_Blackrot | 1000 | 800 | 200 | |
| Apple_Cedarrust | 1000 | 800 | 200 | |
| Apple_scab | 1000 | 800 | 200 | |
Figure 2An example of the dataset augmentation. (a) Original image. (b) Affine transformation. (c) Superimposed Gaussian noise. (d) Flip vertically.
Dataset augmentation process.
| Original data (sheets) | Operations | Final data (sheets) | |
|---|---|---|---|
| The first time | 21,035 | Affine transformation | 42,070 |
| The second time | 42,070 | Superimposed Gaussian noise | 84,140 |
| The third time | 84,140 | Flip vertically | 168,280 |
Figure 3The structure of the improved AlexNet.
The detailed structure of the AlexNet-Inception-V4 network.
| Layer name | Tensor size |
|---|---|
| Input | [3, 227, 227] |
| Conv | [48, 221, 221] |
| Stem | [384, 25, 25] |
| Inception-A | [384, 25, 25] |
| Reduction-A | [1024, 12, 12] |
| Inception-B | [1024, 12, 12] |
| Reduction-B | [1536, 5, 5] |
| Inception-C | [1536, 5, 5] |
| Average pooling | [1536, 1, 1] |
| Fully connected layers | [1536] |
| Output | [4] |
Figure 4The comparison before and after adding dropout. (a) Without dropout. (b) With dropout.
Figure 5Accuracy and loss comparison of the different optimizers. (a) Loss of different optimizers. (b) Accuracy of different optimizers.
Figure 6Accuracy and Loss comparison of different network models. (a) Accuracy comparison. (b) Loss comparison.
Figure 7The comparison of accuracy and loss before and after dataset expansion. (a) Accuracy comparison (b) Loss comparison.
Figure 8The comparison of the accuracy and loss of the training set and the test set.
Figure 9The confusion matrix of four plants (a) Confusion matrix of corn dataset (b) Confusion matrix of tomato dataset (c) Confusion matrix of grape dataset (d) Confusion matrix of apple dataset.
The accuracy, precision, recall, and F1 score for the proposed model.
| Accuracy | Precision | Recall |
| |
|---|---|---|---|---|
| Corn_CercosporaGrayspot |
| 0.940 | 0.937 | 0.938 |
| Corn_Commonrust | 0.955 | 0.907 | 0.913 | 0.910 |
| Corn_healthy | 0.964 | 0.927 | 0.930 | 0.928 |
| Corn_NorthernBlight | 0.960 | 0.927 | 0.900 | 0.917 |
| Tomato_Bacterialspot | 0.948 | 0.905 | 0.887 | 0.896 |
| Tomato_healthy | 0.949 | 0.907 | 0.887 | 0.897 |
| Tomato_Lateblight | 0.936 | 0.870 | 0.873 | 0.872 |
| Tomato_Septorialspot | 0.954 | 0.891 | 0.930 | 0.910 |
| Grape_Blackrot |
| 0.937 | 0.943 | 0.940 |
| Grape_healthy | 0.964 | 0.919 | 0.940 | 0.929 |
| Grape_IsariopsisSpot | 0.953 | 0.923 | 0.883 | 0.903 |
| Grape_Measles |
| 0.944 | 0.947 | 0.945 |
| Apple_Blackrot | 0.960 | 0.929 | 0.910 | 0.919 |
| Apple_Cedarrust | 0.950 | 0.880 | 0.927 | 0.903 |
| Apple_healthy | 0.963 | 0.941 | 0.907 | 0.924 |
| Apple_scab | 0.956 | 0.910 | 0.913 | 0.912 |
Figure 10ROC curve.