| Literature DB >> 36203732 |
Stephen Opoku Oppong1, Frimpong Twum2, James Ben Hayfron-Acquah2, Yaw Marfo Missah2.
Abstract
Computer vision is the science that enables computers and machines to see and perceive image content on a semantic level. It combines concepts, techniques, and ideas from various fields such as digital image processing, pattern matching, artificial intelligence, and computer graphics. A computer vision system is designed to model the human visual system on a functional basis as closely as possible. Deep learning and Convolutional Neural Networks (CNNs) in particular which are biologically inspired have significantly contributed to computer vision studies. This research develops a computer vision system that uses CNNs and handcrafted filters from Log-Gabor filters to identify medicinal plants based on their leaf textural features in an ensemble manner. The system was tested on a dataset developed from the Centre of Plant Medicine Research, Ghana (MyDataset) consisting of forty-nine (49) plant species. Using the concept of transfer learning, ten pretrained networks including Alexnet, GoogLeNet, DenseNet201, Inceptionv3, Mobilenetv2, Restnet18, Resnet50, Resnet101, vgg16, and vgg19 were used as feature extractors. The DenseNet201 architecture resulted with the best outcome of 87% accuracy and GoogLeNet with 79% preforming the worse averaged across six supervised learning algorithms. The proposed model (OTAMNet), created by fusing a Log-Gabor layer into the transition layers of the DenseNet201 architecture achieved 98% accuracy when tested on MyDataset. OTAMNet was tested on other benchmark datasets; Flavia, Swedish Leaf, MD2020, and the Folio dataset. The Flavia dataset achieved 99%, Swedish Leaf 100%, MD2020 99%, and the Folio dataset 97%. A false-positive rate of less than 0.1% was achieved in all cases.Entities:
Mesh:
Year: 2022 PMID: 36203732 PMCID: PMC9532088 DOI: 10.1155/2022/1189509
Source DB: PubMed Journal: Comput Intell Neurosci
Handcrafted features with supervised classifiers.
| Reference | Features | Dataset | Algorithm | Accuracy (%) |
|---|---|---|---|---|
| [ | Texture and shape features | Medicinal plant specimen library of anhui university of traditional Chinese medicine | SVM classifier | 93.3% |
|
| ||||
| [ | Perimeter, a number of vertices, length, width, perimeter and area of hull, colour | Dataset of 24 different plant species having 30 images each from the tropical island of Mauritius | Random forest classifier | 90.1% |
|
| ||||
| [ | Leaf shape and venation structure features | Philippine herbal medicine plants using leaf features | Logistic regression, naïve bayes, K-nearest neighbor (KNN), linear discriminant analysis, classification and regression trees, SVM, and neural networks (NN) | 98.6% |
|
| ||||
| [ | Texture, colour, and shape | Herbal medicinal plants on a dataset containing 50 different species having 500 leaves. | Neural networks | 93.3% |
|
| ||||
| [ | Color, texture and shape feature | Ayurvedic medicinal plant | SVM | 96.66% |
|
| ||||
| [ | Centroid contour curve form signature, a fine-scale margin feature histogram and an interior texture feature histogram | Fisher's iris plant, wheat seed kernels, and 100 plant leaves | Extreme learning machine (ELM) algorithm with K-nearest neighbor, decision tree classifier, support vector machine, naive bayes classifier, and a multilayer perceptron trained with backpropagation algorithm | Iris data set (97%) Seed data set (96%). |
|
| ||||
| [ | Shape, texture, and colour | A total of 3,150 leaf photos from 25 different herbal, fruit, and vegetable species | Support vector machine, K-nearest neighbors, multilayer perceptron, random forest, and decision tree algorithms | 85.82 |
|
| ||||
| [ | 14 features were selected using a chi-square feature selection strategy | Six varieties of medicinal plant leaves | Multilayer perceptron, random forest, logit-boost, basic logistic, and bagging | 99.01% |
|
| ||||
| [ | Morpho-colourimetric parameters Visible (VIS)/Near infrared (NIR) spectral analysis | 20 different Chinese medicinal plants | ANN model | 98.3% |
|
| ||||
| [ | Texture and colour features | Swedish leaf dataset | Multiclass-support vector machine | 93.26%. |
Deep learning models.
| Reference | Algorithm | Dataset | Accuracy (%) |
|---|---|---|---|
| [ | GoogLeNet + linear SVM | 87.34%. | |
| [ | Convolution neural network | 86% | |
| [ | Five-layered convolutional neural network (CNN) | Flavia leaf dataset | 98.22%. |
| Swedish leaf dataset | |||
| [ | CNN-LSTM network with 20 layers | 95.06%. | |
| [ | MobileNetV2 | 98.97 | |
| [ | Dual-path CNN (DP-CNN) | 95.67% | |
| [ | Dual-path CNN model | 14 species of Taiwan's most prevalent trees | 77.1% |
| [ | AlexNet, GoogLeNet, VGG-19, ResNet50, and MobileNetV2 | Leafsnap image dataset | 92.3% |
| [ | 5-Layer CNN architecture | Flavia leaf dataset Swedish leaf dataset | 95.5 98.2 |
| [ | GoogleNet, VGGNet, and AlexNet | LIFECLEF 2015 dataset | 80% |
| [ | Two AlexNets pretrained models | 99.3% | |
| [ | ResNet152 and Inception-ResNetv2 architectures with LBP | Swedish leaf dataset | 99% |
| [ | Seven-layer CNN | Flavia dataset | 94% |
| [ | AlexNet and GoogLeNet | Flavia | 94% |
| Folio | |||
| Swedish leaf dataset | |||
| [ | 17-Layer CNN architecture | 97.9% | |
| [ | VGG19 architecture with a logistic regression classifier | Folio | 96% |
| Flavia | |||
| Swedish leaf datasets | |||
| [ | AousethNet | Mendeley dataset (MD2020 | 99% |
Description of dataset.
|
|
Description of pretrained networks.
| No | Network | Image input size | Depth | No of features | Layer |
|---|---|---|---|---|---|
| 1 | Alexnet | 227-By-227 | 8 | 4096 | fc7 |
| 2 | DenseNet201 | 224-By-224 | 201 | 1920 | avg_pool |
| 3 | Googlenet | 224-By-224 | 22 | 1024 | pool5-7x7_s1 |
| 4 | inceptionv3 | 299-By-299 | 48 | 2048 | avg_pool |
| 5 | mobilenetv2 | 224-By-224 | 53 | 1280 | global_average_pooling2d_1 |
| 6 | resnet18 | 224-By-224 | 18 | 512 | pool5 |
| 7 | resnet50 | 224-By-224 | 50 | 2048 | avg_pool |
| 8 | resnet101 | 224-By-224 | 101 | 2048 | pool5 |
| 9 | vgg16 | 224-By-224 | 16 | 4096 | fc7 |
| 10 | vgg19 | 224-By-224 | 19 | 4096 | fc7 |
Figure 1Proposed CNN model.
Log-gabor parameters.
| Parameter | Value |
|---|---|
| Number of filter scales | 8 |
| Number of filter orientations | 10 |
| Minimum frequency | 3 |
| Scaling between centre frequencies | 2 |
| Filter bandwidth | 0.65 |
| Angular spread of each filter | 1.5 |
Figure 2Feature Extraction Time for Pretrained networks.
Figure 3Accuracy metric for Pretrained networks.
Figure 4F1 score metric for Pretrained networks.
Figure 5FPR score metric for Pretrained networks.
Figure 6PPV score metric for Pretrained networks.
Figure 7TPR score metric for Pretrained networks.
Figure 8Time metric for Pretrained networks.
Figure 9Model accuracy for MyDataset.
Figure 10Model loss for MyDataset.
Metrics for mydataset.
| Plant leaf | ACC | F1 | TPR | FPR | PPV |
|---|---|---|---|---|---|
| 1 | 1 | 1 | 1 | 0 | 1 |
| 2 | 0.99592 | 0.90909 | 1 | 0.00417 | 0.83333 |
| 3 | 0.99796 | 0.94118 | 1 | 0.00207 | 0.88889 |
| 4 | 1 | 1 | 1 | 0 | 1 |
| 5 | 0.99796 | 0.96552 | 1 | 0.0021 | 0.93333 |
| 6 | 1 | 1 | 1 | 0 | 1 |
| 7 | 1 | 1 | 1 | 0 | 1 |
| 8 | 1 | 1 | 1 | 0 | 1 |
| 9 | 1 | 1 | 1 | 0 | 1 |
| 10 | 1 | 1 | 1 | 0 | 1 |
| 11 | 1 | 1 | 1 | 0 | 1 |
| 12 | 1 | 1 | 1 | 0 | 1 |
| 13 | 0.99388 | 0.89655 | 0.8125 | 0 | 1 |
| 14 | 0.99796 | 0.90909 | 0.83333 | 0 | 1 |
| 15 | 1 | 1 | 1 | 0 | 1 |
| 16 | 1 | 1 | 1 | 0 | 1 |
| 17 | 1 | 1 | 1 | 0 | 1 |
| 18 | 1 | 1 | 1 | 0 | 1 |
| 19 | 1 | 1 | 1 | 0 | 1 |
| 20 | 0.99592 | 0.875 | 1 | 0.00414 | 0.77778 |
| 21 | 1 | 1 | 1 | 0 | 1 |
| 22 | 0.99796 | 0.92308 | 0.85714 | 0 | 1 |
| 23 | 1 | 1 | 1 | 0 | 1 |
| 24 | 1 | 1 | 1 | 0 | 1 |
| 25 | 1 | 1 | 1 | 0 | 1 |
| 26 | 1 | 1 | 1 | 0 | 1 |
| 27 | 1 | 1 | 1 | 0 | 1 |
| 28 | 1 | 1 | 1 | 0 | 1 |
| 29 | 1 | 1 | 1 | 0 | 1 |
| 30 | 1 | 1 | 1 | 0 | 1 |
| 31 | 1 | 1 | 1 | 0 | 1 |
| 32 | 0.99796 | 0.94118 | 1 | 0.00207 | 0.88889 |
| 33 | 0.99796 | 0.95652 | 0.91667 | 0 | 1 |
| 34 | 1 | 1 | 1 | 0 | 1 |
| 35 | 1 | 1 | 1 | 0 | 1 |
| 36 | 0.99592 | 0.9 | 1 | 0.00416 | 0.81818 |
| 37 | 1 | 1 | 1 | 0 | 1 |
| 38 | 0.99796 | 0.97143 | 0.94444 | 0 | 1 |
| 39 | 1 | 1 | 1 | 0 | 1 |
| 40 | 1 | 1 | 1 | 0 | 1 |
| 41 | 1 | 1 | 1 | 0 | 1 |
| 42 | 1 | 1 | 1 | 0 | 1 |
| 43 | 1 | 1 | 1 | 0 | 1 |
| 44 | 1 | 1 | 1 | 0 | 1 |
| 45 | 0.99592 | 0.88889 | 0.8 | 0 | 1 |
| 46 | 1 | 1 | 1 | 0 | 1 |
| 47 | 1 | 1 | 1 | 0 | 1 |
| 48 | 1 | 1 | 1 | 0 | 1 |
| 49 | 1 | 1 | 1 | 0 | 1 |
Figure 11Model accuracy on optimizers.
Figure 12Validation accuracy on optimizers.
Figure 13Model loss on optimizers.
Figure 14Validation loss on optimizers.
Metrics based on optimizer.
| Optimizer | Accuracy (%) | Loss |
|---|---|---|
| Adam | 98 | 0.08 |
| RMSProp | 97 | 0.11 |
| SGD | 97 | 0.14 |
Overall statistics.
| Metric | Score |
|---|---|
| ACC | 0.98163 |
| F1 | 0.98117 |
| FPR | 0.00038 |
| PPV | 0.98246 |
| TPR | 0.98294 |
Figure 15Model accuracy for flavia dataset.
Figure 16Model loss for flavia dataset.
Metrics for flavia dataset.
| Plant leaf | ACC | F1 | TPR | FPR | PPV |
|---|---|---|---|---|---|
| 1 | 1 | 1 | 1 | 0 | 1 |
| 2 | 1 | 1 | 1 | 0 | 1 |
| 3 | 1 | 1 | 1 | 0 | 1 |
| 4 | 1 | 1 | 1 | 0 | 1 |
| 5 | 1 | 1 | 1 | 0 | 1 |
| 6 | 0.99738 | 0.9697 | 1 | 0.00273 | 0.94118 |
| 7 | 1 | 1 | 1 | 0 | 1 |
| 8 | 1 | 1 | 1 | 0 | 1 |
| 9 | 1 | 1 | 1 | 0 | 1 |
| 10 | 1 | 1 | 1 | 0 | 1 |
| 11 | 1 | 1 | 1 | 0 | 1 |
| 12 | 1 | 1 | 1 | 0 | 1 |
| 13 | 1 | 1 | 1 | 0 | 1 |
| 14 | 0.99738 | 0.93333 | 0.875 | 0 | 1 |
| 15 | 1 | 1 | 1 | 0 | 1 |
| 16 | 1 | 1 | 1 | 0 | 1 |
| 17 | 1 | 1 | 1 | 0 | 1 |
| 18 | 1 | 1 | 1 | 0 | 1 |
| 19 | 1 | 1 | 1 | 0 | 1 |
| 20 | 1 | 1 | 1 | 0 | 1 |
| 21 | 1 | 1 | 1 | 0 | 1 |
| 22 | 1 | 1 | 1 | 0 | 1 |
| 23 | 1 | 1 | 1 | 0 | 1 |
| 24 | 1 | 1 | 1 | 0 | 1 |
| 25 | 1 | 1 | 1 | 0 | 1 |
| 26 | 1 | 1 | 1 | 0 | 1 |
| 27 | 0.99738 | 0.95652 | 1 | 0.0027 | 0.91667 |
| 28 | 1 | 1 | 1 | 0 | 1 |
| 29 | 1 | 1 | 1 | 0 | 1 |
| 30 | 1 | 1 | 1 | 0 | 1 |
| 31 | 0.99738 | 0.96774 | 0.9375 | 0 | 1 |
| 32 | 1 | 1 | 1 | 0 | 1 |
Figure 17Model accuracy for Swedish leaf dataset.
Figure 18Model loss for Swedish leaf dataset.
Metrics for swedish leaf dataset.
| Plant leaf | ACC | F1 | TPR | FPR | PPV |
|---|---|---|---|---|---|
| 1 | 1 | 1 | 1 | 0 | 1 |
| 2 | 1 | 1 | 1 | 0 | 1 |
| 3 | 1 | 1 | 1 | 0 | 1 |
| 4 | 1 | 1 | 1 | 0 | 1 |
| 5 | 1 | 1 | 1 | 0 | 1 |
| 6 | 1 | 1 | 1 | 0 | 1 |
| 7 | 1 | 1 | 1 | 0 | 1 |
| 8 | 1 | 1 | 1 | 0 | 1 |
| 9 | 1 | 1 | 1 | 0 | 1 |
| 10 | 1 | 1 | 1 | 0 | 1 |
| 11 | 1 | 1 | 1 | 0 | 1 |
| 12 | 1 | 1 | 1 | 0 | 1 |
| 13 | 1 | 1 | 1 | 0 | 1 |
| 14 | 1 | 1 | 1 | 0 | 1 |
| 15 | 1 | 1 | 1 | 0 | 1 |
Figure 19Model accuracy for mendeley dataset.
Figure 20Model loss for mendeley dataset.
Metrics for mendeley dataset.
| Plant leaf | ACC | F1 | TPR | FPR | PPV |
|---|---|---|---|---|---|
| 1 | 1 | 1 | 1 | 0 | 1 |
| 2 | 1 | 1 | 1 | 0 | 1 |
| 3 | 1 | 1 | 1 | 0 | 1 |
| 4 | 1 | 1 | 1 | 0 | 1 |
| 5 | 0.99455 | 0.96154 | 1 | 0.00585 | 0.92593 |
| 6 | 1 | 1 | 1 | 0 | 1 |
| 7 | 1 | 1 | 1 | 0 | 1 |
| 8 | 1 | 1 | 1 | 0 | 1 |
| 9 | 0.99455 | 0.9 | 0.81818 | 0 | 1 |
| 10 | 1 | 1 | 1 | 0 | 1 |
| 11 | 1 | 1 | 1 | 0 | 1 |
| 12 | 1 | 1 | 1 | 0 | 1 |
| 13 | 1 | 1 | 1 | 0 | 1 |
| 14 | 1 | 1 | 1 | 0 | 1 |
| 15 | 0.99183 | 0.91429 | 0.88889 | 0.00287 | 0.94118 |
| 16 | 1 | 1 | 1 | 0 | 1 |
| 17 | 0.99728 | 0.96296 | 1 | 0.00282 | 0.92857 |
| 18 | 1 | 1 | 1 | 0 | 1 |
| 19 | 1 | 1 | 1 | 0 | 1 |
| 20 | 1 | 1 | 1 | 0 | 1 |
| 21 | 1 | 1 | 1 | 0 | 1 |
| 22 | 1 | 1 | 1 | 0 | 1 |
| 23 | 1 | 1 | 1 | 0 | 1 |
| 24 | 1 | 1 | 1 | 0 | 1 |
| 25 | 1 | 1 | 1 | 0 | 1 |
| 26 | 1 | 1 | 1 | 0 | 1 |
| 27 | 1 | 1 | 1 | 0 | 1 |
| 28 | 1 | 1 | 1 | 0 | 1 |
| 29 | 1 | 1 | 1 | 0 | 1 |
| 30 | 0.99455 | 0.8 | 0.8 | 0.00276 | 0.8 |
Figure 21Model accuracy for folio dataset.
Figure 22Model loss for mendeley dataset.
Metrics for folio dataset.
| Plant leaf | ACC | F1 | TPR | FPR | PPV |
|---|---|---|---|---|---|
| 1 | 1 | 1 | 1 | 0 | 1 |
| 2 | 1 | 1 | 1 | 0 | 1 |
| 3 | 0.98438 | 0.83333 | 1 | 0.01626 | 0.71429 |
| 4 | 1 | 1 | 1 | 0 | 1 |
| 5 | 1 | 1 | 1 | 0 | 1 |
| 6 | 1 | 1 | 1 | 0 | 1 |
| 7 | 0.99219 | 0.85714 | 0.75 | 0 | 1 |
| 8 | 1 | 1 | 1 | 0 | 1 |
| 9 | 1 | 1 | 1 | 0 | 1 |
| 10 | 1 | 1 | 1 | 0 | 1 |
| 11 | 1 | 1 | 1 | 0 | 1 |
| 12 | 1 | 1 | 1 | 0 | 1 |
| 13 | 1 | 1 | 1 | 0 | 1 |
| 14 | 1 | 1 | 1 | 0 | 1 |
| 15 | 1 | 1 | 1 | 0 | 1 |
| 16 | 0.98438 | 0.75 | 1 | 0.016 | 0.6 |
| 17 | 1 | 1 | 1 | 0 | 1 |
| 18 | 1 | 1 | 1 | 0 | 1 |
| 19 | 1 | 1 | 1 | 0 | 1 |
| 20 | 1 | 1 | 1 | 0 | 1 |
| 21 | 1 | 1 | 1 | 0 | 1 |
| 22 | 0.99219 | 0.92308 | 0.85714 | 0 | 1 |
| 23 | 1 | 1 | 1 | 0 | 1 |
| 24 | 1 | 1 | 1 | 0 | 1 |
| 25 | 1 | 1 | 1 | 0 | 1 |
| 26 | 0.98438 | 0.5 | 0.33333 | 0 | 1 |
| 27 | 1 | 1 | 1 | 0 | 1 |
| 28 | 1 | 1 | 1 | 0 | 1 |
| 29 | 1 | 1 | 1 | 0 | 1 |
| 30 | 1 | 1 | 1 | 0 | 1 |
| 31 | 1 | 1 | 1 | 0 | 1 |
Figure 23Model accuracy for all dataset.
Figure 24F1 score for all dataset.
Figure 25Fpr for all dataset.
Figure 26PPV for all dataset.
Figure 27Tpr for all dataset.
Figure 28Running time for all dataset.
Comparison with existing systems.
| Source | Method | Dataset | Accuracy (%) |
|---|---|---|---|
| OTAMNet | Log-gabor filter and DenseNet201 | MyDataset | 98 |
| Flavia | |||
| Swedish | |||
| Folio | |||
| MD2020 | |||
| [ | Modified AlexNet | MD2020 | 99 |
| [ | AlexNet, GoogLeNet, VGG-19, ResNet50, and MobileNetV2 | Leafsnap | 92 |
| [ | Binarized Neural Network (BNN) | Swedish leaf | 77 |
| [ | CNN | Flavia | 98 |
| [ | Histogram of oriented gradient (HoG) and deep convolutional neural network | Flavia | 96 |
| [ | VGG19 with LR | Folio | 96 |
| 96 | |||
| 99 | |||
| [ | AlexNet and VGG16 with LDA | Swedish leaf dataset | 99 |
| [ | 17-Layer CNN architecture | LeafSnap | 97 |
| Flavia | |||
| Foliage datasets | |||
| [ | AlexNet and GoogLeNet | Flavia | 94 |
| 99 | |||
| [ | GoogLeNet, VGGNet, and AlexNet | LifeClef 2015 dataset | 80 |
| [ | 26-Layer CNN architecture | BJFU100 dataset | 91 |
| [ | 7-Layer CNN architecture | Flavia dataset | 94 |
| [ | ResNet152 and Inception-ResNetv2 with LBP | Swedish leaf dataset | 99 |