| Literature DB >> 35755682 |
Suiyan Tan1, Jingbin Liu1, Henghui Lu1, Maoyang Lan1, Jie Yu1, Guanzhong Liao1, Yuwei Wang2, Zehua Li3, Long Qi2, Xu Ma2.
Abstract
Recognizing rice seedling growth stages to timely do field operations, such as temperature control, fertilizer, irrigation, cultivation, and disease control, is of great significance of crop management, provision of standard and well-nourished seedlings for mechanical transplanting, and increase of yield. Conventionally, rice seedling growth stage is performed manually by means of visual inspection, which is not only labor-intensive and time-consuming, but also subjective and inefficient on a large-scale field. The application of machine learning algorithms on UAV images offers a high-throughput and non-invasive alternative to manual observations and its applications in agriculture and high-throughput phenotyping are increasing. This paper presented automatic approaches to detect rice seedling of three critical stages, BBCH11, BBCH12, and BBCH13. Both traditional machine learning algorithms and deep learning algorithms were investigated the discriminative ability of the three growth stages. UAV images were captured vertically downward at 3-m height from the field. A dataset consisted of images of three growth stages of rice seedlings for three cultivars, five nursing seedling densities, and different sowing dates. In the traditional machine learning algorithm, histograms of oriented gradients (HOGs) were selected as texture features and combined with the support vector machine (SVM) classifier to recognize and classify three growth stages. The best HOG-SVM model obtained the performance with 84.9, 85.9, 84.9, and 85.4% in accuracy, average precision, average recall, and F1 score, respectively. In the deep learning algorithm, the Efficientnet family and other state-of-art CNN models (VGG16, Resnet50, and Densenet121) were adopted and investigated the performance of three growth stage classifications. EfficientnetB4 achieved the best performance among other CNN models, with 99.47, 99.53, 99.39, and 99.46% in accuracy, average precision, average recall, and F1 score, respectively. Thus, the proposed method could be effective and efficient tool to detect rice seedling growth stages.Entities:
Keywords: SVM; deep learning; growth stage; histograms of oriented gradients; machine learning; rice seedling
Year: 2022 PMID: 35755682 PMCID: PMC9225317 DOI: 10.3389/fpls.2022.914771
Source DB: PubMed Journal: Front Plant Sci ISSN: 1664-462X Impact factor: 6.627
FIGURE 1The schematic diagram of the proposed method.
FIGURE 2Study site and experimental design. (A) Location of the study site; (B) orthomosaic; (C) field tray nursing seedling experiment designs.
Details of RGB images acquisition and the corresponding phenological growth stages of the rice seedlings.
| Inspection date (2021) | 16 March | 19 March | 24 March | |||
|
|
|
| ||||
| Region No. | Growth stage | Images acquired | Growth stage | Images acquired | Growth stage | Images acquired |
| 1. | BBCH11 | 56 | BBCH12 | 83 | BBCH13 | 108 |
| 2. | BBCH11 | 46 | BBCH12 | 60 | BBCH13 | 85 |
| 3. | BBCH11 | 45 | BBCH12 | 92 | BBCH13 | 108 |
| 4. | BBCH11 | 69 | BBCH12 | 98 | BBCH13 | 96 |
| 5. | BBCH11 | 20 | BBCH12 | 32 | BBCH13 | 28 |
| 6. | BBCH11 | 84 | BBCH12 | 120 | BBCH13 | 102 |
| 7. | × | × | BBCH12 | 55 | BBCH13 | 40 |
Detailed information of the datasets.
| The number of images | Growth stages of the seedlings | ||
|
| |||
| BBCH11 | BBCH12 | BBCH13 | |
| Original images | 320 | 540 | 567 |
| Image of 600 × 600 pixels | 1,130 | 3,652 | 4,101 |
| Image of 400 × 400 pixels | 2,814 | 3,545 | 5,739 |
| Image of 300 × 300 pixels | 4,318 | 5,564 | 4,672 |
| Image of 200 × 200 pixels | 8,083 | 18,994 | 25,842 |
| Image of 224 × 224 pixels | 12,357 | 27,830 | 25,393 |
| Image of 100 × 100 pixels | 49,840 | 89,333 | 90,759 |
FIGURE 3Visualization of the extracted HOG features of different seedling growth stages using a cell size of 40 × 40 and block size of 5 × 5 on image of 200 × 200 pixels. Top row, RGB images; bottom row, HOG features of rice seedlings: (A) BBCH11, (B) BBCH12, and (C) BBCH13.
Parameters of the EfficientnetB0 network.
| Stage( | Operator ( | Resolution ( | Channels ( | Layers ( |
| 1 | Conv3 × 3 | 224 × 224 | 32 | 1 |
| 2 | MBConv1, k3 × 3 | 112 × 112 | 16 | 1 |
| 3 | MBConv6, k3 × 3 | 112 × 112 | 24 | 2 |
| 4 | MBConv6, k5 × 5 | 56 × 56 | 40 | 2 |
| 5 | MBConv6, k3 × 3 | 28 × 28 | 80 | 3 |
| 6 | MBConv6, k5 × 5 | 14 × 14 | 112 | 3 |
| 7 | MBConv6, k5 × 5 | 14 × 14 | 192 | 4 |
| 8 | MBConv6, k3 × 3 | 7 × 7 | 320 | 1 |
| 9 | Conv1 × 1 & Pooling & FC | 7 × 7 | 1,280 | 1 |
FIGURE 4Diagram of the EfficientnetB4 used to detect rice seedling growth stages.
FIGURE 5Evaluation performance on the validation sets for SVMs trained on the HOG features as a function of SVM kernels and input image sizes: (A) Accuracy, (B) average precision, (C) average recall, and (D) F1 score. Error bars show the standard deviation across the SVMs.
FIGURE 6Evaluation performance on the validation sets for SVMs trained on the HOG features as a function of cell size c and block size b: (A) Accuracy, (B) average precision, (C) average recall, and (D) F1 score.
Confusion matrix for SVM using HOG feature with cell size of 16 and block size of 2 evaluated on the test set.
| Predicted | |||||
|
| |||||
| BBCH11 | BBCH12 | BBCH13 | Recall | ||
| Observed | BBCH11 | 172 | 20 | 9 | 85.57% |
| BBCH12 | 5 | 177 | 19 | 88.06% | |
| BBCH13 | 0 | 38 | 163 | 81.09% | |
| Precision | 97.18% | 75.32% | 85.34% | 84.91% | |
The lower right cell shows the accuracy.
Evaluation performance on test sets for HOG-SVM classifiers with different numbers of training images in each growth stage.
| Input image size(pixels) | Number of training images in each growth stage | Accuracy | Average precision | Average recall | F1 score | Time of training(min) | Time of testing(sec) |
| 400 × 400 | 1,000 | 84.9% | 85.9% | 84.9% | 85.4% | 3.03 | 12.374 |
| 400 × 400 | 1,500 | 81.0% | 83.3% | 81.1% | 82.2% | 4.16 | 35.741 |
| 400 × 400 | 2,000 | 81.4% | 83.0% | 81.4% | 82.2% | 7.24 | 52.320 |
| 400 × 400 | 2,500 | 79.8% | 80.8% | 79.8% | 80.3% | 11.6 | 76.156 |
Evaluation performance on the validation sets for EfficientnetB0-B7.
| Models | Accuracy | Average precision | Average recall | F1 score | Time of training (min) | Time of test (sec) |
| EfficientnetB0 | 97.67 | 97.54 | 97.71 | 97.62 | 106.7 | 32 |
| EfficientnetB1 | 97.52 | 97.54 | 97.75 | 97.64 | 40.9 | 12 |
| EfficientnetB2 | 97.61 | 97.61 | 97.79 | 97.70 | 36.5 | 12 |
| EfficientnetB3 | 98.43 | 98.41 | 98.58 | 98.50 | 50.0 | 15 |
| EfficientnetB4 | 99.47 | 99.53 | 99.39 | 99.46 | 62.5 | 19 |
| EfficientnetB5 | 98.78 | 99.09 | 99.01 | 99.05 | 83.2 | 24 |
| EfficientnetB6 | 98.28 | 98.60 | 98.46 | 98.53 | 135.0 | 40 |
| EfficientnetB7 | 98.72 | 98.85 | 98.74 | 98.79 | 220.9 | 66 |
Confusion matrix for EfficientnetB4 evaluated on the test set.
| Predicted | |||||
|
| |||||
| BBCH11 | BBCH12 | BBCH13 | Recall | ||
| Observed | BBCH11 | 563 | 0 | 0 | 100 |
| BBCH12 | 3 | 696 | 10 | 98.17% | |
| BBCH13 | 0 | 0 | 1,148 | 100% | |
| Precision | 99.47% | 100% | 99.14% | 99.47% | |
The lower right cell shows the accuracy.
FIGURE 7Precision–Recall curves of different deep learning models: (A) EfficientnetB0-B7 (B) State-of-art CNN models.
Evaluation performance on the validation sets for different CNN models.
| Accuracy | Average precision | Average recall | F1 score | Total training time (min) | Total test time (sec) | |
| EfficientnetB4 | 99.47% | 99.53% | 99.39% | 99.46% | 62.5 | 19 |
| Densenet121 | 99.06% | 98.79% | 99.11% | 98.95% | 114.2 | 35 |
| Resnet50 | 98.97% | 98.74% | 98.92% | 98.83% | 104.2 | 30 |
| VGG16 | 94.84% | 94.55% | 94.83% | 94.69% | 116.7 | 33 |
Evaluation performance on test sets for EfficientnetB4 with different numbers of training images in each growth stage.
| Model | Number of training images in each growth stage | Accuracy | Average precision | Average recall | F1 score | Time of training (min) | Time of testing (sec) |
| EfficientnetB4 | 1,000 | 98.17 | 98.20 | 98.17 | 98.18 | 15.8 | 5 |
| EfficientnetB4 | 2,000 | 98.75 | 98.78 | 98.75 | 98.77 | 30.8 | 9 |
| EfficientnetB4 | 2,500 | 99.15 | 99.15 | 99.16 | 99.16 | 45.0 | 13 |
FIGURE 8Effects of different input image size on HOG-SVM classification.