| Literature DB >> 30152813 |
Timm Schoening1, Kevin Köser1, Jens Greinert1,2.
Abstract
Optical imaging is a common technique in ocean research. Diving robots, towed cameras, drop-cameras and TV-guided sampling gear: all produce image data of the underwater environment. Technological advances like 4K cameras, autonomous robots, high-capacity batteries and LED lighting now allow systematic optical monitoring at large spatial scale and shorter time but with increased data volume and velocity. Volume and velocity are further increased by growing fleets and emerging swarms of autonomous vehicles creating big data sets in parallel. This generates a need for automated data processing to harvest maximum information. Systematic data analysis benefits from calibrated, geo-referenced data with clear metadata description, particularly for machine vision and machine learning. Hence, the expensive data acquisition must be documented, data should be curated as soon as possible, backed up and made publicly available. Here, we present a workflow towards sustainable marine image analysis. We describe guidelines for data acquisition, curation and management and apply it to the use case of a multi-terabyte deep-sea data set acquired by an autonomous underwater vehicle.Entities:
Year: 2018 PMID: 30152813 PMCID: PMC6111891 DOI: 10.1038/sdata.2018.181
Source DB: PubMed Journal: Sci Data ISSN: 2052-4463 Impact factor: 6.444
Figure 1Schematic overview of the proposed image data workflow from acquisition through curation and management.
Various robots (autonomous underwater vehicles (AUVs), landers, remotely operated vehicles (ROVs), towed platforms) create stacks of imagery (a) and metadata tables (b). Erroneous metadata values (here marked in red) and corrupt imagery (e.g. black images where the flash did not fire) might occur. Metadata are attached to the image data, image processing is applied and corrupt and erroneous data are flagged and filtered out (c). The resulting curated data set is the quality controlled data product that is suitable for publication and analysis. Metadata and image data are stored in suitable databases (public or private). Image data items should be linked to their corresponding metadata at archiving. The individual steps from pre-cruise planning to publication are discussed in the text. For a specific use case, see Fig. 3.
Figure 2Example images from the presented use case image data set.
Panel (a) shows a raw image taken by the camera onboard AUV Abyss from ca. 10 m altitude. The object in the middle is a stationary lander[26] which was deployed independently of the AUV dives for environmental measurements. Panel (b) shows the effect of lens un-distortion. Black boxes show areas of 6, 3 and 2 m2 (top to bottom), corresponding to average footprints of other optical image acquisition gear (i.e. AUV, towed camera, ROV), computed for their usual operational altitude and field of view. Images in panels (c) and (e) are further un-distorted examples, taken at altitudes of 7.5 m and 4.5 m. Images in (d) and (f) are the results of a color normalization, applied to (c) and (e). Two zoom-ins (marked by the dashed, white box) show an anemone surrounded by poly-metallic nodules (d) and a sea star, close to a decades-old plow track that extend over the entire image (f, parallel linear structures).
Figure 3The data workflow as applied to the AUV use case.
The AUV Abyss created metadata files (a) and stacks of up to 50,000 images (b) per dive. Meta- and image data were fused by time code (d). Un-distortion was applied (e), erroneous data were removed (f). Raw metadata are stored in OSIS (c). Raw and curated imagery is managed with ProxSys (g, h). Curated image data are made publicly available: in PANGAEA for long-term archival (i - by duplication) and in BIIGLE for manual annotation (j - by link). OSIS links to the image data in BIIGLE (k). Subsequent image analysis, enabled by the curated data are color normalization (i), mosaicking (n), mineral resource mapping (o), and automated event detection and classification within individual images (p), using manual annotations from BIIGLE (m) and machine learning.
Metadata fields to be stored alongside each image to geo-reference each pixel of an image.
| These represent the best-case scenario where all parameters are easily measurable. We propose to use these exact tag terms to enable data interchange between image analysis softwares. The chosen tags are derived from the field names used in the world data center PANGAEA for arbitrary marine data (changed to lowercase and without blanks and special characters to streamline automated processing). All lengths measurements in mm. | |
|---|---|
| SUB_datetime | A date-time-stamp in the format "YYYYMMDD hh:mm:ss.sss" |
| SUB_latitude | Latitude position of the camera platform in decimal degrees |
| SUB_longitude | Longitude position of the camera platform in decimal degrees |
| SUB_distance | Altitude of the camera platform above ground |
| SUB_heading | Direction of travel along the x-axis, 0=North, 90=East |
| SUB_forwardvelocity | Speed of the camera platform along the x-axis |
| SUB_yawangle | Yaw angle (rotation around z-axis, see[ |
| SUB_pitchangle | Pitch angle (rotation around y-axis, see[ |
| SUB_rollangle | Roll angle (rotation around x-axis, see[ |
| CAM_model | Manufacturer and type of camera |
| CAM_id | Machine-readable camera identifier for multi-camera systems |
| CAM_position | Relative offsets of camera center in camera platform coordinates |
| CAM_orientation | Relative orientation of the camera to the camera platform |
| CAM_refraction | Refraction data: glass port type, glass thickness, refractive index |
| CAM_alignment | Port offset and normal in camera coordinates |
| CAM_lensmodel | Manufacturer and type of objective lens |
| CAM_focallength | Focal length in mm |
| CAM_fnumber | Objective lens aperture information |
| ENV_temperaturewater | Optional temperature parameter |
| ENV_absorption | Optional parameter for the absorption coefficients of the water |
| ENV_scattering | Optional parameter for the volume scattering function |
| ENV_refractiveindex | Optional refractive index parameter of the water around the camera platform |
| REF_laserdistances | Optional parameter if laser points are used for scaling |