| Literature DB >> 33313554 |
Yu Jiang1,2,3, Changying Li2,3.
Abstract
Plant phenotyping has been recognized as a bottleneck for improving the efficiency of breeding programs, understanding plant-environment interactions, and managing agricultural systems. In the past five years, imaging approaches have shown great potential for high-throughput plant phenotyping, resulting in more attention paid to imaging-based plant phenotyping. With this increased amount of image data, it has become urgent to develop robust analytical tools that can extract phenotypic traits accurately and rapidly. The goal of this review is to provide a comprehensive overview of the latest studies using deep convolutional neural networks (CNNs) in plant phenotyping applications. We specifically review the use of various CNN architecture for plant stress evaluation, plant development, and postharvest quality assessment. We systematically organize the studies based on technical developments resulting from imaging classification, object detection, and image segmentation, thereby identifying state-of-the-art solutions for certain phenotyping applications. Finally, we provide several directions for future research in the use of CNN architecture for plant phenotyping purposes.Entities:
Year: 2020 PMID: 33313554 PMCID: PMC7706326 DOI: 10.34133/2020/4152816
Source DB: PubMed Journal: Plant Phenomics ISSN: 2643-6515
Figure 1Diagram of the pathway of imaging-based plant phenotyping.
Summary of major CNN architecture developed for image classification, object detection, and semantic and instance segmentation.
| Model | Vision task | Key concept | Source code (third-party implementation) |
|---|---|---|---|
| AlexNet | Image classification | A five-layer CNN architecture |
|
|
| |||
| ZFNet | Image classification | Feature visualization for model improvement |
|
|
| |||
| VGGNet | Image classification | Small-sized (3 by 3) convolutional filters to increase the depth of CNNs (up to 19 layers) |
|
|
| |||
| Inception family | Image classification | Inception modules for increasing the width of CNNs and therefore the capability of feature representation |
|
|
| |||
| ResNet family | Image classification | Residual representation and skip connection scheme to enable the training of very deep CNNs (up to 1000 layers) |
|
|
| |||
| DenseNet | Image classification | Dense block modules to substantially decrease the number of model parameters (therefore computational cost) and strengthen feature propagation (therefore feature learning capability) |
|
|
| |||
| NASNet | Image classification | Reinforcement learning on a small dataset to find optimal convolutional cells that are used to build a CNN architecture for a large dataset |
|
|
| |||
| RCNN family | Object detection | A two-stage framework to generate regions of interest (ROIs) and then predict the class label and calculate the bounding box coordinates for each ROI |
|
|
| |||
| YOLO family | Object detection | A one-stage framework to regress both class labels and bounding box coordinates for each grid cell on the last feature map |
|
|
| |||
| SSD | Object detection | A one-stage framework to regress class labels and bounding box coordinates for anchors in each grid cell on feature maps extracted from different convolution layers (thus different resolutions) |
|
|
| |||
| RetinaNet | Object detection | A one-stage framework to use focal loss that is a new loss function to solve the foreground-background class imbalance problem |
|
|
| |||
| FCN | Semantic segmentation | Fully convolutional architecture to train and predict classes at the pixel level in an end-to-end manner for semantic segmentation |
|
|
| |||
| U-Net | Semantic segmentation | An encoder-decoder architecture for semantic segmentation |
|
|
| |||
| DeepLab family | Semantic segmentation | Atrous convolution operation to simultaneously increase receptive field and reduce the computation complexity to improve the segmentation accuracy; fully connected conditional random field (CRF) as a postprocessing method to improve the segmentation accuracy |
|
|
| |||
| Mask RCNN | Instance segmentation | Masking head with ROI align operation on top of the Faster RCNN model to significantly improve segmentation accuracy |
|
Note: ∗source code provided by original authors.
Figure 2Diagrams of CNN architecture mechanisms for image classification, object detection, and semantic and instance segmentation.
Figure 3Key concept and results of xPlNet for plant stress detection: (a) diagram of the developed xPlNet for calculating the explanation map for a given image; (b) visualization results using different methods for an image containing a stressed leaf. (a) and (b) were reproduced using figures from [55, 56], respectively.
Figure 4Diagrams of key concepts for using CNNs for plant/organ detection, counting, and localization.
Summary of CNN-based data analysis approach in imaging-based plant phenotyping.
| Phenotyping category | Phenotyping task | Main processing approach | Particular improvement strategy | References |
|---|---|---|---|---|
| Plant stress | Stress detection and classification | Image classification | NA | [ |
| Sliding window | [ | |||
| Explainable visualization | [ | |||
| Advanced imaging | [ | |||
| Synthetic data augmentation | [ | |||
| Object detection | NA | [ | ||
|
| ||||
| Plant development | Plant lodging | Image classification | NA | [ |
| Canopy morphology measurement | Object detection | NA | [ | |
| Semantic segmentation | NA | [ | ||
| Leaf morphology measurement | Instance segmentation | NA | [ | |
| Characterization of plant growth pattern | Combination of CNN and other DL methods | NA | [ | |
|
| ||||
| Plant development | Counting plant/plant organs in still images | Regression | NA | [ |
| Synthetic data augmentation | [ | |||
| Multiscale and multimodal data fusion | [ | |||
| Nonsupervised learning mode | [ | |||
| Explainable visualization | [ | |||
| Image classification | NA | [ | ||
| Object detection | NA | [ | ||
| Sliding window | [ | |||
| Synthetic data augmentation | [ | |||
| Semantic segmentation | NA | [ | ||
| Sliding window | [ | |||
| Instance segmentation | NA | [ | ||
| Synthetic data augmentation | [ | |||
|
| ||||
| Plant development | Counting plant/plant organs in image sequences and videos | Object detection | 2D orthoimage reconstruction | [ |
| 3D structure reconstruction | [ | |||
| Video tracking | [ | |||
| Semantic segmentation | Movement encoding | [ | ||
|
| ||||
| Plant development | Counting root tips | Regression | NA | [ |
| Root system architecture segmentation | Semantic segmentation | NA | [ | |
| Inpainting for oversegmentation correction | [ | |||
| Advanced imaging | [ | |||
| Synthetic data augmentation | [ | |||
|
| ||||
| Postharvest quality | Fruit chemical composition measurement | Regression | NA | [ |
| Fruit defect detection | Image classification | NA | [ | |
| Advanced imaging | [ | |||
| Sliding window | [ | |||
| Fruit defect quantification | Semantic segmentation | NA | [ | |
| Advanced imaging | [ | |||