| Literature DB >> 32626607 |
Alexander E White1,2, Rebecca B Dikow1, Makinnon Baugh3, Abigail Jenkins3, Paul B Frandsen1,3.
Abstract
PREMISE: Digitized images of herbarium specimens are highly diverse with many potential sources of visual noise and bias. The systematic removal of noise and minimization of bias must be achieved in order to generate biological insights based on the plants rather than the digitization and mounting practices involved. Here, we develop a workflow and data set of high-resolution image masks to segment plant tissues in herbarium specimen images and remove background pixels using deep learning. METHODS ANDEntities:
Keywords: U‐Net; deep learning; digitized herbarium specimens; ferns; machine learning; semantic segmentation
Year: 2020 PMID: 32626607 PMCID: PMC7328659 DOI: 10.1002/aps3.11352
Source DB: PubMed Journal: Appl Plant Sci ISSN: 2168-0450 Impact factor: 1.936
FIGURE 1Herbarium sheets and associated masks made available in this study. (A) Four example digitized herbarium sheets from the U.S. National Herbarium at the National Museum of Natural History (Washington, D.C., USA). (B) The same four sheets shown as high‐resolution masks. A total of 400 masks were generated using the methods described in the text and were used to train a deep neural net to automatically segment plant tissues from herbarium specimens. Species names for each image clockwise from top left: Rumohra adiantiformis (G. Forst.) Ching, Thelypteris kunthii (Desv.) C. V. Morton (synonym Christella kunthii), Asplenium peruvianum var. insulare (C. V. Morton) D. D. Palmer, Thelypteris palustris Schott.
FIGURE 2Workflow outlining the automatic and manual steps in generating the image masks and training the U‐Net. High‐resolution JPEG (.jpg) files were exported from the Smithsonian Digital Asset Management System to the High‐Performance Computing Cluster where we ran the segmentation Python code. Outputs from this step were edited in Adobe Photoshop to remove label and color palette before running the postprocessing code (binarize and blur tools) that produced the final ground‐truth masks. These ground‐truth masks were then used as training data for the U‐Net model.
FIGURE 3A comparison of high‐resolution original images, ground‐truth masks, and U‐Net‐predicted mask outputs. (A) Two example original images. (B) Ground‐truth masks. (C) Mask outputs predicted by U‐Net (Sørensen–Dice coefficient = 0.95). Note that the predicted masks are all resized to 256 × 256 pixels to maximize downstream model training efficiency regardless of image input size. The square output predictions crop rectangular inputs. Species names from top: Callistopteris apiifolia (Presl) Copel., Ceradenia capillaris (Desv.) L. E. Bishop.
Model performance for individual fern families.
| Family | No. of validation images | Sørensen–Dice coefficient |
|---|---|---|
| Gleicheniaceae | 4 | 0.972 |
| Lygodiaceae | 4 | 0.966 |
| Hymenophyllaceae | 13 | 0.922 |
| Equisetaceae | 5 | 0.917 |
| Ophioglossaceae | 4 | 0.959 |
| Marattiaceae | 3 | 0.971 |
| Psilotaceae | 1 | 0.906 |
| Osmundaceae | 1 | 0.963 |
| Schizaeaceae | 2 | 0.882 |
| Anemiaceae | 2 | 0.952 |
| Cyatheaceae | 12 | 0.968 |
| Polypodiaceae | 5 | 0.948 |
| Dryopteridaceae | 4 | 0.948 |
| Pteridaceae | 6 | 0.963 |
| Tectariaceae | 1 | 0.916 |
| Aspleniaceae | 3 | 0.963 |
| Lindsaeaceae | 1 | 0.936 |
| Blechnaceae | 1 | 0.963 |
| Thelypteridaceae | 4 | 0.954 |
| Athyriaceae | 1 | 0.861 |
| Salviniaceae | 1 | 0.960 |
| Dicksoniaceae | 1 | 0.958 |
| Marsileaceae | 1 | 0.942 |