| Literature DB >> 36247386 |
Eric C Orenstein1, Sakina-Dorothée Ayata1,2, Frédéric Maps3,4, Érica C Becker5, Fabio Benedetti6, Tristan Biard7, Thibault de Garidel-Thoron8, Jeffrey S Ellen9, Filippo Ferrario3,4,10, Sarah L C Giering11, Tamar Guy-Haim12, Laura Hoebeke13, Morten Hvitfeldt Iversen14, Thomas Kiørboe15, Jean-François Lalonde16, Arancha Lana17, Martin Laviale18, Fabien Lombard1, Tom Lorimer19, Séverine Martini20, Albin Meyer18, Klas Ove Möller21, Barbara Niehoff14, Mark D Ohman9, Cédric Pradalier22, Jean-Baptiste Romagnan23, Simon-Martin Schröder24, Virginie Sonnet25, Heidi M Sosik26, Lars S Stemmann1, Michiel Stock13, Tuba Terbiyik-Kurt27, Nerea Valcárcel-Pérez28, Laure Vilgrain1, Guillaume Wacquet29, Anya M Waite30, Jean-Olivier Irisson1.
Abstract
Plankton imaging systems supported by automated classification and analysis have improved ecologists' ability to observe aquatic ecosystems. Today, we are on the cusp of reliably tracking plankton populations with a suite of lab-based and in situ tools, collecting imaging data at unprecedentedly fine spatial and temporal scales. But these data have potential well beyond examining the abundances of different taxa; the individual images themselves contain a wealth of information on functional traits. Here, we outline traits that could be measured from image data, suggest machine learning and computer vision approaches to extract functional trait information from the images, and discuss promising avenues for novel studies. The approaches we discuss are data agnostic and are broadly applicable to imagery of other aquatic or terrestrial organisms.Entities:
Year: 2022 PMID: 36247386 PMCID: PMC9543351 DOI: 10.1002/lno.12101
Source DB: PubMed Journal: Limnol Oceanogr ISSN: 0024-3590 Impact factor: 5.019
Fig. 1Plankton functional traits that can be estimated from images, following the unified typology of Martini et al. (2021). Trait types along the y‐axis follow the order of the “Plankton traits from images” section. Measured traits, ones that can be quantified solely from images, are in capital letters. Inferred traits, which require additional information beyond raw pixels, are written in bold text.
Fig. 2Example of plankton images on which traits can be identified. (a–h) Diatoms, (i–v) copepods, (w–δ) other taxa. (a–c) Chains of Chaetoceros spp. of different sizes (Scripps Pier Cam [SPC]); note the long spines on (c). (d) Sexual stage of Guinardia flaccida (Imaging FlowCytobot [IFCB]). (e) Dinoflagellate consuming a diatom chain (Guinardia delicatula) by external digestion in a feeding veil (pallium) (IFCB). (f) Guinardia delicatula infected with parasite (first arrow) or as an empty frustule (second arrow) [IFCB]. (g) Ditylum brightwellii cell dividing (IFCB). (h) Coscinodiscophycidae (centric diatoms) containing various amounts of pigments (Planktoscope). (i) Nauplius stage of a crustacean (ZooScan), (j–m) calanoid copepods (Underwater Vision Profiler 5), note the full gut (arrow) and active posture with antennae deployed on (j), the pigmented (dark) body parts on (j–l), the lipid sac (arrow) and resting posture, with antennae along the body on (l), and the curved antennae (arrow) associated with a jump of the copepod on (m). (n) Immature (top) and mature (bottom, with visible oocytes—arrow) female of Calanus hyperboreus (Lightframe On‐sight Key species Investigation [LOKI]). (o) Gaetanus brevispinus displaying many sensory setae on its antennae and a well visible gut (arrow) (LOKI). (p) Another copepod with well visible setae and two egg sacs (arrows) (SPC). (q) Copepod associated with (possibly feeding on) a marine snow particle (ZooGlider). (r) Microsetella sp. displaying many spines and intense coloration, likely from its gut content (Planktoscope). (s) Calanoid copepod with parasite dinoflagellates (arrow) (ZooCAM). (t) Male (with geniculate antennae—arrow, left) and female (with bulging genital segment—arrow, right) of Centropages sp. (ZooCAM). (u) Oncaea mating [SPC]. (v) Empty copepod carcass or molt (ZooScan). (w) Doliolid budding (ISIIS). (x) Salp with an amphipod inside (arrow) (UVP). (y) Transparent Doliolid (SPC). (z) A few solitary Rhizaria, family Aulospheridae (ZooGlider), to be contrasted with (δ). (α) Foraminifera with long cell extensions (UVP). (β) Pteropod (dark) with part of its mucus net deployed (gray). (ɣ) Ctenophore, family Mertensiidae, with very long fishing tentacles deployed (ISIIS). (δ) A colonial Rhizaria, order Collodaria (ZooGlider).
Definitions of a few computational terms, highlighting important, but subtle, differences.
| Computer vision (CV) | A broad subfield of computer science dedicated to using a computer to interpret images and video sequences. |
| Machine learning (ML) | A set of statistical approaches that attempt to discern patterns in data, either automatically or based on explicit human instructions. |
| Supervised ML | ML techniques that teach a computer to recognize patterns using a set of expert‐curated examples, such as annotated images. |
| Unsupervised ML | ML methods that attempt to group data together without human intervention. Clustering algorithms are a common example. Their performance is often difficult to evaluate. |
| Training set | A collection of data annotated by human experts for teaching a computer how to interpret information. Building the labeled dataset is the most time‐consuming and critical part of an ML workflow. |
| Validation set | A separate human labeled dataset used to evaluate a trained system. These data are entirely independent of the training set and should represent conditions the system might encounter in the field. Also referred to as |
| Feature‐based learning | ML algorithms that operate on a reduced, hand‐engineered feature space. Each data point is cast as a vector of measurements and used to tune a set of parameters that dictate how the model works. |
| Deep neural networks (DNNs) | A type of |
| Transfer learning | A shortcut for training DNNs by repurposing a network originally trained for a different task. |
Fig. 3Workflow diagram for a computer vision approach targeting a specific functional trait extracted from plankton image data. In this example, a group decides to target egg‐bearing copepods with a UNet segmentation model. A human annotator selects ovigerous tissue from a copepod image and outputs the mask in the COCO data format. The model is then trained and evaluated using the mAP (Table 2; Supporting Information S1). Note that this workflow is not specific to plankton and could also be used for other types of organisms.
Common evaluation metrics for automated classifiers. See Supporting Information for more detail and description.
| Accuracy (acc) |
| The number of true positives in a class returned by an automated classifier over the total number of correct labels in the class. |
| Precision ( |
| The ratio of the number of correct labels over the total number of labels assigned to that class. |
| Recall ( |
| The proportion of positive samples in a class that are correctly classified. |
| F1‐score |
| The harmonic mean of precision and recall that summarizes model performance in a single metric scaling between 0 and 1. |
| Average precision (AP) |
| The average of the precision over different levels of the recall. |
| Mean average precision (mAP) |
| The mean AP over all classes. It is a summary statistic that describes how a model does over all classes. |
| Intersection over union (IoU) |
|
|
Fig. 4Examples of several techniques for trait extraction from zooplankton images. The hypothetical use case is examining ovigerous copepods imaged by the Scripps Plankton Camera system. The top panel is a non‐egg bearing copepod. The bottom panel is an individual carrying an egg‐sac. (a) Automated classifiers could be trained to add a semantic descriptor to the taxonomic class. (b) Object detection finds the organism and desired trait. (c) Segmentation algorithms classify the pixels as belonging to the organism or the trait. (d) Regression estimates the percentage of pixels that represent the trait. (e) Keypoint/pose estimation finds body nodes (red dots) and connects them (yellow lines) to estimate orientation or appendage extension.
Fig. 5Examples of several techniques for trait extraction from phytoplankton. The hypothetical use case is examining parasitized diatom chains imaged by the Imaging FlowCytobot. The top panel is a healthy Guinardia delicatula. The bottom panel is a parasitized chain. (a) Automated classifiers could be trained to add a semantic descriptor to the taxonomic class. (b) Object detection finds the entire chain, chloroplasts, and parasites. (c) Segmentation algorithms classify individual pixels as belonging to the organism, chloroplasts, or parasites. (d) Regression estimates the amount of an image that corresponds to the organelle/parasite biovolume.