| Literature DB >> 35406899 |
Nikos Tsiknakis1, Elisavet Savvidaki2, Georgios C Manikis1, Panagiota Gotsiou3, Ilektra Remoundou3, Kostas Marias1,4, Eleftherios Alissandrakis2, Nikolas Vidakis4.
Abstract
Pollen identification is an important task for the botanical certification of honey. It is performed via thorough microscopic examination of the pollen present in honey; a process called melissopalynology. However, manual examination of the images is hard, time-consuming and subject to inter- and intra-observer variability. In this study, we investigated the applicability of deep learning models for the classification of pollen-grain images into 20 pollen types, based on the Cretan Pollen Dataset. In particular, we applied transfer and ensemble learning methods to achieve an accuracy of 97.5%, a sensitivity of 96.9%, a precision of 97%, an F1 score of 96.89% and an AUC of 0.9995. However, in a preliminary case study, when we applied the best-performing model on honey-based pollen-grain images, we found that it performed poorly; only 0.02 better than random guessing (i.e., an AUC of 0.52). This indicates that the model should be further fine-tuned on honey-based pollen-grain images to increase its effectiveness on such data.Entities:
Keywords: classification; deep learning; ensemble; honey certification; melissopalynology; pollen grain; transfer learning
Year: 2022 PMID: 35406899 PMCID: PMC9002917 DOI: 10.3390/plants11070919
Source DB: PubMed Journal: Plants (Basel) ISSN: 2223-7747
Figure 1Data histogram across classes.
Figure 2Mosaic of images of all pollen types: 1 Thymbra; 2 Erica; 3 Castanea; 4 Eucalyptus; 5 Myrtus; 6 Ceratonia; 7 Urginea; 8 Vitis; 9 Origanum; 10 Satureja; 11 Pinus; 12 Calicotome; 13 Salvia; 14 Sinapis; 15 Ferula; 16 Asphodelus; 17 Oxalis; 18 Pistacia; 19 Ebenus; 20 Olea.
Figure 3Histogram of each subset of the data.
Augmentations used in the study.
| Augmentation Method | Hyperparameters | Probability |
|---|---|---|
| Gaussian Blurring | Sigma [0, 0.3] | 30% |
| Linear Contrast Adjustment | Alpha [0.75, 1.25] | 30% |
| Brightness Multiplication | Multiplication factor [0.7, 1.3] | 30% |
| Rotation | Degrees [−180, 180] | 100% |
| Translation in x Plane | Translation percentage [−0.2, 0.2] | 100% |
| Translation in y Plane | Translation percentage [−0.2, 0.2] | 100% |
| Vertical Flipping | - | 50% |
| Horizontal Flipping | - | 50% |
Classification part of the network. The input and output sizes of the GAP layer, as well as the input size of the first dense layer, depend on the backbone convolutional network used.
| Input Size | Output Size | Activation Function | |
|---|---|---|---|
| Global Average Pooling 2D | - | - | - |
| Dense Layer | - | 1024 | ReLU |
| Droput of 50% | 1024 | 1024 | - |
| Dense Layer | 1024 | 512 | ReLU |
| Droput of 50% | 512 | 512 | - |
| Dense Layer | 512 | 256 | ReLU |
| Droput of 50% | 256 | 256 | - |
| Dense Layer | 256 | 128 | ReLU |
| Droput of 50% | 1024 | 1024 | - |
| Dense Layer | 128 | 20 | Softmax |
Example of the ensemble strategies for a binary classification task. The same procedure applies for a 20-class task.
| Models | Prediction Probability | Prediction |
|---|---|---|
| Model 1 | [0.8, 0.2] | Class 0 |
| Model 2 | [0.55, 0.45] | Class 0 |
| Model 3 | [0.1, 0.9] | Class 1 |
| Soft Voting Ensemble | [0.8 + 0.55 + 0.1, 0.2 + 0.45 + 0.9]/3 = [0.483, 0.517] | Class 1 |
| Hard Voting Ensemble | Maximum occurrence [0,0,1] | Class 0 |
Performance metrics for each model on the test set, averaged across all classes.
| Macro | Weighted | ||||||||
|---|---|---|---|---|---|---|---|---|---|
| ACC | Pre | Sen | F1 | AUC | Pre | Sen | F1 | AUC | |
|
| 0.975161 | 0.970762 | 0.966647 | 0.967991 | NA | 0.976031 | 0.975161 | 0.975231 | NA |
|
| 0.975161 | 0.970042 | 0.969219 | 0.968880 |
| 0.976306 | 0.975161 | 0.975334 |
|
|
| 0.972678 | 0.969362 | 0.966809 | 0.967251 | NA | 0.973837 | 0.972678 | 0.972838 | NA |
|
| 0.973671 | 0.969781 | 0.966628 | 0.967443 | 0.999437 | 0.974688 | 0.973671 | 0.973803 | 0.999338 |
|
| 0.957278 | 0.952362 | 0.947819 | 0.947840 | NA | 0.959952 | 0.957278 | 0.957202 | NA |
|
| 0.966716 | 0.965132 | 0.963203 | 0.962795 | 0.999358 | 0.969911 | 0.966716 | 0.967410 | 0.999204 |
|
| 0.964729 | 0.959518 | 0.956076 | 0.956596 | NA | 0.966231 | 0.964729 | 0.964742 | NA |
|
| 0.974168 | 0.967764 | 0.970974 | 0.968863 | 0.999177 | 0.975065 | 0.974168 | 0.974326 | 0.999204 |
|
| 0.959265 | 0.958602 | 0.949879 | 0.951970 | NA | 0.962451 | 0.959265 | 0.959344 | NA |
|
| 0.972181 | 0.971409 | 0.969076 | 0.969541 | NA | 0.973656 | 0.972181 | 0.972402 | NA |
|
| 0.974168 |
| 0.968699 | 0.969395 | 0.999222 | 0.975805 | 0.974168 | 0.974393 | 0.999170 |
|
| 0.969697 | 0.967619 | 0.964934 | 0.965034 | NA | 0.972198 | 0.969697 | 0.970201 | NA |
|
| 0.971187 | 0.968954 | 0.966567 | 0.966683 | 0.999464 | 0.973259 | 0.971187 | 0.971571 | 0.999475 |
|
| 0.966716 | 0.963960 | 0.961913 | 0.961237 | 0.998892 | 0.970550 | 0.966716 | 0.967457 | 0.998980 |
|
| 0.965723 | 0.960852 | 0.954880 | 0.956783 | NA | 0.966856 | 0.965723 | 0.965620 | NA |
|
| 0.976155 | 0.969923 | 0.970366 | 0.969455 | NA | 0.977386 | 0.976155 | 0.976399 | NA |
|
|
| 0.969540 | 0.971659 |
| 0.999387 |
|
|
| 0.999454 |
|
| 0.974168 | 0.967077 |
| 0.968744 | 0.998931 | 0.975470 | 0.974168 | 0.974411 | 0.999008 |
|
| 0.967213 | 0.960367 | 0.952631 | 0.954955 | NA | 0.968760 | 0.967213 | 0.967131 | NA |
|
| 0.971684 | 0.964913 | 0.963993 | 0.963454 | 0.999097 | 0.973144 | 0.971684 | 0.971921 | 0.999184 |
|
| 0.964729 | 0.960787 | 0.961660 | 0.960547 | 0.998633 | 0.966119 | 0.964729 | 0.964959 | 0.998587 |
|
| 0.952310 | 0.953253 | 0.952506 | 0.950253 | 0.998199 | 0.958644 | 0.952310 | 0.953534 | 0.997607 |
|
| 0.958271 | 0.950490 | 0.955365 | 0.951001 | 0.998360 | 0.962226 | 0.958271 | 0.959258 | 0.998233 |
|
| 0.961749 | 0.950759 | 0.953611 | 0.950483 | 0.998096 | 0.964355 | 0.961749 | 0.962129 | 0.998363 |
Individual models (short name for ensemble naming): Inception (i); ResNet (r); Inception–ResNet (ir); Xception (x). Ensemble strategies: soft or hard voting. The naming of the ensemble models is based on the schema “ens_MODELS_VOTING” (e.g., ens_i_r_soft means Inception and ResNet ensembles based on soft voting). NA refers to Not Applicable.
Performance metrics for each model on the test set regarding the Thymbra class.
| Sensitivity | Specificity | Precision | Accuracy | F1 | AUC | |
|---|---|---|---|---|---|---|
|
| 0.917808 | 0.998969 | 0.971014 | 0.996026 | 0.943662 | NA |
|
| 0.931507 | 0.998969 | 0.971429 | 0.996523 | 0.951049 |
|
|
| 0.917808 | 0.998969 | 0.971014 | 0.996026 | 0.943662 | NA |
|
| 0.931507 | 0.998969 | 0.971429 | 0.996523 | 0.951049 | 0.999364 |
|
| 0.780822 |
|
| 0.992052 | 0.876923 | NA |
|
| 0.917808 |
|
|
|
| 0.998757 |
|
| 0.835616 | 0.998969 | 0.968254 | 0.993045 | 0.897059 | NA |
|
| 0.931507 | 0.997938 | 0.944444 | 0.995529 | 0.937931 | 0.999244 |
|
| 0.821918 |
|
| 0.993542 | 0.902256 | NA |
|
| 0.904110 | 0.998454 | 0.956522 | 0.995032 | 0.929577 | NA |
|
| 0.931507 | 0.998969 | 0.971429 | 0.996523 | 0.951049 | 0.995220 |
|
| 0.917808 | 0.998969 | 0.971014 | 0.996026 | 0.943662 | NA |
|
| 0.904110 | 0.999485 | 0.985075 | 0.996026 | 0.942857 | 0.998913 |
|
| 0.904110 |
|
| 0.996523 | 0.949640 | 0.995629 |
|
| 0.863014 | 0.998454 | 0.954545 | 0.993542 | 0.906475 | NA |
|
| 0.931507 | 0.998454 | 0.957746 | 0.996026 | 0.944444 | NA |
|
| 0.931507 | 0.998969 | 0.971429 | 0.996523 | 0.951049 | 0.999477 |
|
|
| 0.996907 | 0.920000 | 0.995032 | 0.932432 | 0.995523 |
|
| 0.821918 | 0.998969 | 0.967742 | 0.992548 | 0.888889 | NA |
|
| 0.904110 | 0.998969 | 0.970588 | 0.995529 | 0.936170 | 0.998687 |
|
| 0.931507 | 0.995361 | 0.883117 | 0.993045 | 0.906667 | 0.994549 |
|
| 0.849315 |
|
| 0.994536 | 0.918519 | 0.993913 |
|
| 0.863014 | 0.997938 | 0.940299 | 0.993045 | 0.900000 | 0.997522 |
|
| 0.904110 | 0.996907 | 0.916667 | 0.993542 | 0.910345 | 0.995763 |
Figure 4Receiver operating characteristic curve for Thymbra class.
Figure 5Receiver operating characteristic curve for soft voting ensemble of all base models.
Performance results of soft voting ensemble of all base models across all classes.
| Sensitivity | Specificity | Precision | Accuracy | F1 | AUC | |
|---|---|---|---|---|---|---|
|
| 0.931507 | 0.998969 | 0.971429 | 0.996523 | 0.951049 | 0.999569 |
|
| 1.000000 | 0.998439 | 0.968085 | 0.998510 | 0.983784 | 1.000000 |
|
| 1.000000 | 0.998950 | 0.981982 | 0.999006 | 0.990909 | 1.000000 |
|
| 0.941176 | 0.998444 | 0.963855 | 0.996026 | 0.952381 | 0.999713 |
|
| 0.989822 | 1.000000 | 1.000000 | 0.998013 | 0.994885 | 0.999991 |
|
| 0.960000 | 0.995925 | 0.857143 | 0.995032 | 0.905660 | 0.998839 |
|
| 1.000000 | 1.000000 | 1.000000 | 1.000000 | 1.000000 | 1.000000 |
|
| 0.962963 | 0.995208 | 0.935252 | 0.993045 | 0.948905 | 0.999101 |
|
| 0.941176 | 0.999481 | 0.987654 | 0.997019 | 0.963855 | 0.995973 |
|
| 0.972222 | 0.998988 | 0.945946 | 0.998510 | 0.958904 | 0.999930 |
|
| 1.000000 | 1.000000 | 1.000000 | 1.000000 | 1.000000 | 1.000000 |
|
| 0.946309 | 0.997854 | 0.972414 | 0.994039 | 0.959184 | 0.999622 |
|
| 1.000000 | 1.000000 | 1.000000 | 1.000000 | 1.000000 | 1.000000 |
|
| 1.000000 | 0.993730 | 0.891892 | 0.994039 | 0.942857 | 0.999609 |
|
| 0.975610 | 1.000000 | 1.000000 | 0.999503 | 0.987654 | 0.999975 |
|
| 1.000000 | 0.999499 | 0.944444 | 0.999503 | 0.971429 | 1.000000 |
|
| 1.000000 | 0.999485 | 0.985915 | 0.999503 | 0.992908 | 1.000000 |
|
| 0.882353 | 1.000000 | 1.000000 | 0.999006 | 0.937500 | 0.999882 |
|
| 0.909091 | 1.000000 | 1.000000 | 0.999503 | 0.952381 | 0.999273 |
|
| 0.972152 | 0.998764 | 0.994819 | 0.993542 | 0.983355 | 0.999368 |
Figure 6Confusion matrix for soft voting ensemble of all base models.
Comparison table between other studies and ours.
| Ref. | Method | Dataset | AUC | Sensitivity | Precision | Accuracy | F1 Score |
|---|---|---|---|---|---|---|---|
| Manikis et al. [ | Hand-crafted Features + ML | 546 images | - | 88.16% | 88.60% | 88.24% | 87.79% |
| Battiato et al. [ | CNN | Pollen23E | - | - | - | 89.63% | 88.97% |
| Sevillano et al. [ | CNN + LD | Pollen23E | - | 99.64% | 94.77% | 93.22% | 96.69% |
| Astolfi et al. [ | CNN | Pollen73S 2523 images | - | 95.7% | 95.7% | 95.8% | 96.4% |
| Our study | CNN | CPD 4034 | 0.9995 | 96.9% | 97% | 97.5% | 96.89% |
Figure 7Histogram of honey-based dataset.
Figure 8Confusion matrix of soft voting ensemble of all models on the honey-based dataset.
Figure 9Pollen-grain images of pollen types with similar morphological characteristics.