| Literature DB >> 33800565 |
Mohamed Shaban1, Reem Salim2, Hadil Abu Khalifeh2, Adel Khelifi2, Ahmed Shalaby3, Shady El-Mashad4, Ali Mahmoud3, Mohammed Ghazal2, Ayman El-Baz3.
Abstract
Oil leaks onto water surfaces from big tankers, ships, and pipeline cracks cause considerable damage and harm to the marine environment. Synthetic Aperture Radar (SAR) images provide an approximate representation for target scenes, including sea and land surfaces, ships, oil spills, and look-alikes. Detection and segmentation of oil spills from SAR images are crucial to aid in leak cleanups and protecting the environment. This paper introduces a two-stage deep-learning framework for the identification of oil spill occurrences based on a highly unbalanced dataset. The first stage classifies patches based on the percentage of oil spill pixels using a novel 23-layer Convolutional Neural Network. In contrast, the second stage performs semantic segmentation using a five-stage U-Net structure. The generalized Dice loss is minimized to account for the reduced oil spill representation in the patches. The results of this study are very promising and provide a comparable improved precision and Dice score compared to related work.Entities:
Keywords: Synthetic Aperture Radar (SAR); deep learning; oil spill
Year: 2021 PMID: 33800565 PMCID: PMC8036558 DOI: 10.3390/s21072351
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1(a) SAR image depicting sea, oil spill, look-alikes, ships, and land; (b) corresponding three-dimensional masks.
Classes of instances and respective labels.
| Class | 3-D Masks | 1-D Labels | Pixel Counts |
|---|---|---|---|
| Sea | Black | 0 | 225,071,583 |
| Oil Spill | Cyan | 1 | 3,188,485 |
| Look-Alike | Red | 2 | 13,585,532 |
| Ship | Brown | 3 | 87,287 |
| Land | Green | 4 | 9,942,113 |
Figure 2(a) Before applying the Frost filter. (b) After applying the Frost filter.
Figure 3Proposed deep-learning framework.
Layers of the Proposed CNN.
| Layer | Kernel Size | No. of Kernels | Kernel Features | No. of Layers |
|---|---|---|---|---|
| Input Layer | 64 × 64 × 3 | 1 | True Color | 1 |
| Convolutional/ReLU Layers | 11 × 11 | 32 | Same Padding | 4 |
| Pooling Layer | 2 × 2 | 32 | No Padding | 1 |
| Convolutional/ReLU Layers | 9 × 9 | 64 | Same Padding | 4 |
| Pooling Layer | 2 × 2 | 64 | No Padding | 1 |
| Convolutional/ReLU Layers | 7 × 7 | 128 | Same Padding | 4 |
| Pooling Layer | 2 × 2 | 128 | No Padding | 1 |
| Fully Connected/ReLU Layers | 128, 64, 32, 16 | 1 | 4 | |
| Fully Connected Layer | 2 | 1 | 1 | |
| SoftMax Layer | 1 | 1 | ||
| Classification Layer | 1 | 1 |
Figure 4Five-stage U-Net Architecture for the Semantic Segmentation of Oil Spills.
First-Stage CNN Performance.
| Measure | 5-Fold (Mean) | 5-Fold (Deviation) | 10-Fold (Mean) | 10-Fold (Deviation) |
|---|---|---|---|---|
| Training Accuracy | 1 | 8.1 × 10−5 | 1 | 2.1 × 10−4 |
| Validation Accuracy | 0.99 | 1.2 × 10−4 | 0.99 | 2.8 × 10−4 |
| Weighted Kappa Score | 0.99 | 2.3 × 10−4 | 0.99 | 5.6 × 10−4 |
| Sensitivity | 0.99 | 2.2 × 10−4 | 0.99 | 6.1 × 10−4 |
| Specificity | 0.99 | 2.2 × 10−4 | 0.99 | 2.3 × 10−4 |
| AUC | 0.99 | 0.0016 | 0.99 | 0.0021 |
Second-Stage U-Net Performance on Training Dataset.
| Measure | Accuracy | Precision | Recall | Dice |
|---|---|---|---|---|
|
| 0.9 | 0.91 | 0.88 | 0.89 |
Second-Stage U-Net Performance on Testing Dataset.
| Measure | Accuracy | Precision | Recall | Dice |
|---|---|---|---|---|
|
| 0.92 | 0.84 | 0.76 | 0.8 |
Comparison between the proposed framework and direct application of semantic segmentation.
| Measure | Proposed Framework | U-Net | SegNet |
|---|---|---|---|
|
| 92% | 95% | 96% |
|
| 76% | 76% | 66% |
|
| 84% | 27% | 26% |
|
| 80% | 40% | 38% |
Proposed framework versus variations of the proposed segmentation network.
| Measure | Proposed Framework | Patch Size 32 × 32 | Patch Size 128 × 128 | Recall Loss Optimization | Jaccard Loss Minimization |
|---|---|---|---|---|---|
|
| 92% | 80% | 94% | 92% | 92% |
|
| 76% | 88% | 70% | 76% | 76% |
|
| 84% | 63% | 81% | 83% | 84% |
|
| 80% | 74% | 75% | 79% | 80% |
Comparative study between the proposed framework and the-state-of-the-art methods.
| Measure | Proposed Framework | Hidalgo et al. [ | Zeng et al. [ |
|---|---|---|---|
|
| 92% | 99% | 94% |
|
| 76% | 86.8% | 83.5% |
|
| 84% | 65% | 85.7% |
|
| 80% | 71% | 84.6% |
Figure 5Testing data ground truth and predicted masks such that (a,c,e,g) ground truth labels; (b,d) predicted masks with more than 90% accuracy; (f) predicted mask with more than 70% accuracy; (h) predicted mask with slightly over 50% accuracy.