| Literature DB >> 35281624 |
Abdelmalek Bouguettaya1, Hafed Zarzour2, Ahmed Kechida1, Amine Mohammed Taberkit1.
Abstract
During the last few years, Unmanned Aerial Vehicles (UAVs) technologies are widely used to improve agriculture productivity while reducing drudgery, inspection time, and crop management cost. Moreover, they are able to cover large areas in a matter of a few minutes. Due to the impressive technological advancement, UAV-based remote sensing technologies are increasingly used to collect valuable data that could be used to achieve many precision agriculture applications, including crop/plant classification. In order to process these data accurately, we need powerful tools and algorithms such as Deep Learning approaches. Recently, Convolutional Neural Network (CNN) has emerged as a powerful tool for image processing tasks achieving remarkable results making it the state-of-the-art technique for vision applications. In the present study, we reviewed the recent CNN-based methods applied to the UAV-based remote sensing image analysis for crop/plant classification to help researchers and farmers to decide what algorithms they should use accordingly to their studied crops and the used hardware. Fusing different UAV-based data and deep learning approaches have emerged as a powerful tool to classify different crop types accurately. The readers of the present review could acquire the most challenging issues facing researchers to classify different crop types from UAV imagery and their potential solutions to improve the performance of deep learning-based algorithms.Entities:
Keywords: Convolutional neural network; Crop classification; Deep learning; Deep neural network; UAV; Unmanned aerial vehicle
Year: 2022 PMID: 35281624 PMCID: PMC8898032 DOI: 10.1007/s00521-022-07104-9
Source DB: PubMed Journal: Neural Comput Appl ISSN: 0941-0643 Impact factor: 5.102
Advantages and drawbacks of the existing remote-sensing platforms for agriculture
| Remote sensing technologies | Advantages | Drawbacks |
|---|---|---|
| Satellite | Very low spatial resolution | |
| Very large area coverage | Cloud sensitivity | |
| Very high spectral resolution | High cost | |
| Data not available all time | ||
| Manned aircraft | Cloud sensitivity | |
| Large area coverage | High cost | |
| High spectral resolution | Low spatial resolution | |
| Affected by weather conditions | ||
| Unmanned aerial vehicle | Low cost | Medium area coverage |
| High spatial/spectral resolution | Low endurance | |
| Data available all time whenever we want it | Affected by weather conditions | |
| Not sensitive to clouds | Require more time than satellite to cover very large areas | |
| Accurate position | Taught flying laws and regulations | |
| Difficult areas accessibility | ||
| Unmanned/manned ground vehicle + other on-ground technologies | Very low area coverage | |
| Very high spatial resolution | Direct impact on the field | |
| Require very long time to cover even small areas |
Fig. 1CNN architecture
Fig. 2Performance evolution of different Deep CNN architectures for Image net classification challenge
Fig. 3Examples of different camera sensors a RGB, b Multispectral, c Hyperspectral, d FLIR Vue Pro R, and e Phoenix LiDAR
Fig. 4Example of wheat detection framework from UAV imagery.
Different object-based crop/plant classification methods
| Ref. | Method | Backbone | Plant types | Sensor | GSD (cm) | Altitude (m) | Average Precision (%) | Precision (%) | Recall (%) | F1-Score (%) | Inference time (ms) | FPS |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| [ | Faster R-CNN | Inception-v2 | Banana plant | RGB | 1.78 | 40 | 99.3 | 96.4 | 97.82 | |||
| 2.03 | 50 | 97.9 | 85.1 | 91.05 | ||||||||
| 2.54 | 60 | 98.5 | 75.8 | 85.67 | ||||||||
| 40 + 50 | 98.3 | 99 | 98.64 | |||||||||
| 40 + 50 + 60 | 97.9 | 98.6 | 98.24 | |||||||||
| [ | RetinaNet | ResNet | Ornamental plants | RGB | 73.41 | |||||||
| YOLOv3 | DarkNet-53 | 79.85 | ||||||||||
| [ | YOLOv2 | Mango fruits | RGB | 1.5 − 2 | 86.4 | 96.1 | 89 | 92.41 | 80 | 40 | ||
| [ | MangoYOLO | / | Mango fruits | RGB | 2 | 98.3 | 96.8 | 15 | 14 | |||
| YOLOv2 | DarkNet-19 | 95.9 | 93.3 | 20 | ||||||||
| YOLOv2-tiny | / | 95.3 | 91.7 | 10 | ||||||||
| YOLOv3 | DarkNet-53 | 96.7 | 95.1 | 25 | ||||||||
| SSD | VGG-16 | 98.3 | 95.9 | 70 | ||||||||
| Faster R-CNN | VGG | 95.3 | 94.5 | 67 | ||||||||
| Faster R-CNN | ZF | 95 | 93.9 | 37 | ||||||||
| [ | Faster R-CNN | ResNet-50 | Trees | RGB | 0.82 | 20 − 40 | 82.48 | 163 | ||||
| YOLOv3 | DarkNet-53 | 85.88 | 26 | |||||||||
| RetinaNet | FPN | 92.64 | 67 | |||||||||
| [ | YOLOv4 | Wheatears | RGB | 0.01 − 0.06 | 1.2 − 3 | 62.75 | 88.23 | 57 | ||||
| Modified YOLOv4 | 77.81 | 96.71 | 72 | |||||||||
| [ | YOLOv3 | DarkNet-53 | Citrus trees | Multi-spectral | 75 | 99.9 | 99.7 | 99.79 | ||||
| [ | CNN + Classifier (with refinement) | Citrus trees | Multi-spectral | 104 | 94.59 | 97.94 | 96.24 | |||||
| CNN + Classifier (without refinement) | 65 | 98 | 78 | |||||||||
| [ | RetinaNet | Citrus trees | Multi-spectral | 12.9 | 120 | 62 | 92 | 74 | ||||
| Faster R-CNN | 86 | 39 | 54 | |||||||||
| Proposed approach | VGG-16 | 95 | 96 | 95 | 40 (8 stages) | 25 | ||||||
| [ | Faster R-CNN | Apple trees | RGB | 91.1 | 94.1 | 92.5 |
Fig. 5Example of semantic segmentation from UAV imagery using U-Net architecture [52].
Different pixel-based crop/plant classification methods
| Ref. | Data types | GSD (cm) | Altitude (m) | Method | Backbone | Target crops | Overall Accuracy (%) | Precision (%) | Recall (%) | Kappa | F1-Score (%) | Time (s) |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| [ | RGB | 5.3 | 230 | SegNet | VGG-16 | Rice paddy, Rice lodging | 83.10 | 69.06 | 75.43 | 101 | ||
| RGB+ExG | 71.03 | 89.64 | 79.26 | 108 | ||||||||
| RGB+ExGR | 73.96 | 83.36 | 78.38 | 109 | ||||||||
| RGB+ExG+ExGR | 87.66 | 57.38 | 69.36 | 106 | ||||||||
| RGB | FCN | AlexNet | 84.73 | 82.43 | 83.56 | 59 | ||||||
| RGB+ExG | 84.92 | 80.85 | 82.84 | 65 | ||||||||
| RGB+ExGR | 77.02 | 88.80 | 82.49 | 66 | ||||||||
| RGB+ExG+ExGR | 82.02 | 84.44 | 83.21 | 72 | ||||||||
| RGB | 5.7 | 200 | SegNet | VGG-16 | Rice paddy, Rice lodging | 87.55 | 38.65 | 53.63 | 99 | |||
| RGB+ExG | 57.06 | 67.50 | 61.84 | 109 | ||||||||
| RGB+ExGR | 81.35 | 58.59 | 68.12 | 107 | ||||||||
| RGB+ExG+ExGR | 82.47 | 29.07 | 42.99 | 113 | ||||||||
| RGB | FCN | AlexNet | 99.12 | 39.59 | 56.58 | 57 | ||||||
| RGB+ExG | 95.03 | 59.18 | 72.94 | 67 | ||||||||
| RGB+ExGR | 95.32 | 66.39 | 78.27 | 68 | ||||||||
| RGB+ExG+ExGR | 93.19 | 67.03 | 77.97 | 71 | ||||||||
| [ | RGB | 6.4 | 150 | FCN | Sunflower | 78 | ||||||
| SegNet | 79 | |||||||||||
| Improved SegNet | VGG-16 | 81.95 | ||||||||||
| Multispectral | 8.1 | FCN | 77.45 | |||||||||
| SegNet | 78.25 | |||||||||||
| Improved SegNet | VGG-16 | 80.5 | ||||||||||
| Fusion (RGB + FNIR) | 9.5 | FCN | 81.55 | |||||||||
| SegNet | 82.65 | |||||||||||
| Improved SegNet | VGG-16 | 86.55 | ||||||||||
| [ | RGB | 9.5 | FCRN-MTL | Full-grown citrus trees | 98.8 | 99.2 | 98.4 | 98.8 | ||||
| Citrus tree seedlings | 56.6 | 47.8 | 65.4 | 55.23 | ||||||||
| [ | RGB | 20 | Deep convolutional Encoder-decoder network | Fig Plant | 93.84 | 92 | 95.05 | 93.50 | ||||
| SegNet-basic | 93.82 | 93.49 | 93.22 | 93.35 | ||||||||
| [ | RGB | 20 | U-Net | Corn | 99.4 | |||||||
| [ | RGB | 10 | FCN | Beet | 49.69 | 26.25 | 32.15 | |||||
| SegNet | 80.24 | 67.4 | 71.79 | |||||||||
| CR-Hough-SLIC | 86.58 | 85.67 | 85.40 | |||||||||
| CRowNet | 90.37 | 90.56 | 90.39 | |||||||||
| CR-Hough-SLIC | Maize | 85.14 | 97.74 | 82.13 | ||||||||
| CRowNet | 84.57 | 80.93 | 82.5 | |||||||||
| [ | RGB | 2.21 | SegNet | VGG-16 | Rice, Corn | 89.44 | ||||||
| FCN | AlexNet | 88.48 | ||||||||||
| [ | Multispectral | U-Net | VGG-16 | Sugar beet | 99 | |||||||
| SegNet | 98 | |||||||||||
| [ | RGB | U-Net | Apple tree crown | 97.1 | 84.2 | 84.5 | 84.2 | |||||
| [ | RGB | 0.37, 0.56, 0.74 | 20 | U-Net | Purple rapeseed leaves | 91.56 | ||||||
| [ | RGB | 3 | 100 | U-Net | Pinus radiata, Ulex europaeus | 85.5 | ||||||
| [ | RGB | 1.7–2 | Mask R-CNN | Potato | 99.72 | 82.50 | 90.3 | |||||
| Lettuce | 100 | 95.43 | 97.7 | |||||||||
| [ | RGB | 1 | 40–60 | Mask Scoring R-CNN | Maize V5 growth stage | 95.8 | 82.8 | |||||
| Multispectral | 2.5 | Maize V4 growth stage | 82.7 | 79.9 | ||||||||
| [ | RGB | 4 | 120 | DeepLabv3+ | ResNet-18 | Amazonian palms | 86 | 88 | 87 | |||
| [ | RGB | DeepLabv3+ | Mauritia flexuosa | 98.036 | 96.688 | 95.616 | 96.14 | |||||
| U-Net1 | 95.973 | 91.381 | 92.632 | 92 | ||||||||
| U-Net2 | 97.682 | 94.858 | 95.953 | 95.4 | ||||||||
| U-Net3 | 96.843 | 92.534 | 94.886 | 93.7 | ||||||||
| U-Net4 | 97.512 | 95.166 | 95.028 | 95.1 | ||||||||
| [ | Hyperspectral | 46.3 | 500 | CNN-CRF | WHU-Hi-LongKou: 6 crop types | 98.91 | 0.9857 | |||||
| 10.9 | 250 | WHU-Hi-HanChuan: 7 crop types | 93.95 | 0.9290 | ||||||||
| 4.3 | 100 | WHU-Hi-HongHu: 17 crop types | 93.74 | 0.9217 | ||||||||
| [ | RGB | 0.47, 0.90, 1.43, and 1.76 | FDN-92 | 12 plants | 87 | |||||||
| FDN-29 | 84 | |||||||||||
| FDN-17 | 86 | |||||||||||
| Inception-V1 | 80 | |||||||||||
| Inception-V2 | 79 | |||||||||||
| Inception-V3 | 77 | |||||||||||
| ResNet-17 | 73 | |||||||||||
| ResNet-50 | 74 | |||||||||||
| ResNet-101 | 76 | |||||||||||
| DenseNet-21 | 82 | |||||||||||
| DenseNet-36 | 80 | |||||||||||
| DenseNet-121 | 82 | |||||||||||
| [ | Hyperspectral | CNN | 19 crop types | 88.62 | 0.8557 | |||||||
| CNN-CRF | 91.79 | 0.8957 | ||||||||||
| [ | RGB | 25 | 2D-CNN | Highland Kimchi cabbage, Cabbage, Potato | 86.56 | |||||||
| [ | RGB | 3 | DNN | VGG-16 | Maize, Bananas, Legumes | 86 | 86 | 86 | 0.82 | 86 | ||
| [ | RGB + Multi-spectral | 100 | LeNet | LeNet | Corn | 67.7 | 46.4 | |||||
| 180 | 86.8 | 81.8 | ||||||||||
| 100-180 | 72.6 | 54.5 | ||||||||||
| 100 | 47.7 | 12.8 | ||||||||||
| [ | RGB | 5 | Hybrid CNN-HistNN | 22 crops | 90 | |||||||
| [ | UAV | ANN | Peanut, Maize, Honeysuckle, Tree | 78.53 | 0.73 | |||||||
| Fused | 85.72 | 0.81 |
UAV-based datasets
| Ref. | Dataset | Data type | Number of images | Image size | Altitude/GSD (m) | Crop type | Acquisition time |
|---|---|---|---|---|---|---|---|
| WHU-Hi-HongHu | 550 | 500 / 0.463 | Corn, cotton, sesame, broad-leaf soybean, narrow-leaf soybean, and rice | July 17, 2018, from 13:49 to 14:37 | |||
| [ | WHU-Hi-HanChuan | Hyperspectral | – | 1217 | 250 / 0.109 | Strawberry, cowpea, soybean, sorghum, water spinach, watermelon, and greens | June 17, 2016, from 17:57 to 18:46 |
| WHU-Hi-HongHu | 940 | 100 / 0.043 | 17 crop types, including cotton, rape, and cabbage | November 20, 2017, from 16:23 to 17:37 | |||
| [ | Weedmap (000) | Multispectral | 10,196 | 5995 | 10 / 1.04 | Sugar beet & weed | 18 September 2017, 9:18–40 a.m. |
| Weedmap (001) | 4867 | 10 / 0.94 | |||||
| Weedmap (002) | 6403 | 10 / 0.96 | |||||
| Weedmap (003) | 5470 | 10 / 0.99 | |||||
| Weedmap (004) | 4319 | 10 / 1.07 | |||||
| Weedmap (005) | 7221 | 10 / 0.85 | 5–18 May 2017, around 12:00 p.m. | ||||
| Weedmap (006) | 5601 | 10 / 1.18 | |||||
| Weedmap (007) | 6074 | 10 / 0.83 | |||||
| [ | VOAI | RGB | 2293 (No data augmentation) | 224 | 20 / 0.47 | 12 tree species | – |
| 21,863 (with data augmentation) | 30 / 0.90 | ||||||
| 40 / 1.43 | |||||||
| 50 / 1.76 | |||||||
| [ | WeedNet | Multispectral | 465 | – | 2 / – | Sugar beet & weed | – |
Evaluation metrics
| Metric | Formula | Description |
|---|---|---|
| Accuracy | Indicates the percentage of true predictions among | |
| all predictions. | ||
| Precision | Determines how well a model is in predicting positive labels. | |
| Recall (sensitivity) | Measures the percentage of true positives successfully | |
| detected by a model. | ||
| F1-score (F-measure) | Represents the harmonic mean of Precision and Recall rates | |
| Kappa coefficient | Measures inter-annotator agreement. | |
| FPS | Determines the detection speed. |
Fig. 6Comparison of different algorithms to classify various crop types a True-color image, b Ground-truth image, c SVM, d FNEA-OO, e SVRFMC, f Benchmark CNN, and g CNNCRF [128].
Fig. 7Examples of object-based (left) [4] and pixel-based (right) [21] crop classification.