| Literature DB >> 30216353 |
Víctor Sevillano1, José L Aznarte2.
Abstract
In palynology, the visual classification of pollen grains from different species is a hard task which is usually tackled by human operators using microscopes. Its complete automatization would save a high quantity of resources and provide valuable improvements especially for allergy-related information systems, but also for other application fields as paleoclimate reconstruction, quality control of honey based products, collection of evidences in criminal investigations or fabric dating and tracking. This paper presents three state-of-the-art deep learning classification methods applied to the recently published POLEN23E image dataset. The three methods make use of convolutional neural networks: the first one is strictly based on the idea of transfer learning, the second one is based on feature extraction and the third one represents a hybrid approach, combining transfer learning and feature extraction. The results from the three methods are indeed very good, reaching over 97% correct classification rates in images not previously seen by the models, where other authors reported around 70.Entities:
Mesh:
Year: 2018 PMID: 30216353 PMCID: PMC6138340 DOI: 10.1371/journal.pone.0201807
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Fig 1Sample images for each pollen type.
Fig 2Convolutional neural network architecture and operation.
Image based on a similar figure published in [26].
Fig 3Cross-validation schematic for setups A and C.
Fig 4Cross-validation schematic for setup B.
Average correct classification rate over the test set for each setup.
Results from [6] are computed from their confusion matrices. In parentheses under the values, standard deviation.
| CCR (%) | precision | recall | F-score | |
|---|---|---|---|---|
| CST+BOW (from [ | 68.5714 | 0.3988 | 0.8203 | 0.5366 |
| Human (from [ | 63.5710 | 0.3030 | 0.6279 | 0.4087 |
| Setup A (FE+LD) | 96.6247 | 0.9366 | 0.9955 | 0.9592 |
| Setup B (TL) | 96.1529 | 0.9275 | 0.9949 | 0.9541 |
| Setup C (TL+FE+LD) | 97.2273 | 0.9477 | 0.9964 | 0.9669 |
Fig 5Correct classification rate for human operators and the best model reported by [6], together with the three proposed deep learning setups.
Fig 6Confusion matrix for the test set in setup A.
Fig 7Confusion matrix for the test set in setup B.
Fig 8Confusion matrix for the test set in setup C.
Measures for computational complexity.
| Prediction speed | Training time | |
|---|---|---|
| Setup A (FE+LD) | ∼140 obs/sec | 8.69 min |
| Setup B (TL) | ∼155 obs/sec | 16.49 min |
| Setup C (TL+FE+LD) | ∼170 obs/sec | 16.61 min |
Fig 9Distribution of the correct classification rate for the test set in setup C, by pollen type.
Types: A) Anadenanthera colubrina, B) Arecaceae, C) Arrabidaea, D) Cecropia pachystachya, E) Chromolaena laevigata, F) Combretum discolor, G) Croton urucurana, H) Dipteryx alata, I) Eucalyptus, J) Faramea, K) Hyptis, L) Mabea fistulifera, M) Matayba guianensis, N) Mimosa somnians, O) Myrcia, P) Protium heptaphyllum, Q) Qualea multiflora, R) Schinus terebinthifolius, S) Senegalia plumosa, T) Serjania laruotteana, U) Syagrus, V) Tridax procumbens, W) Urochloa decumbens.