| Literature DB >> 31991872 |
Anastasios Stamoulakatos1, Javier Cardona1, Chris McCaig1, David Murray2, Hein Filius2, Robert Atkinson1, Xavier Bellekens1, Craig Michie1, Ivan Andonovic1, Pavlos Lazaridis3, Andrew Hamilton1, Md Moinul Hossain4, Gaetano Di Caterina1, Christos Tachtatzis1.
Abstract
Regulatory requirements for sub-sea oil and gas operators mandates the frequent inspection of pipeline assets to ensure that their degradation and damage are maintained at acceptable levels. The inspection process is usually sub-contracted to surveyors who utilize sub-sea Remotely Operated Vehicles (ROVs), launched from a surface vessel and piloted over the pipeline. ROVs capture data from various sensors/instruments which are subsequently reviewed and interpreted by human operators, creating a log of event annotations; a slow, labor-intensive and costly process. The paper presents an automatic image annotation framework that identifies/classifies key events of interest in the video footage viz. exposure, burial, field joints, anodes, and free spans. The reported methodology utilizes transfer learning with a Deep Convolutional Neural Network (ResNet-50), fine-tuned on real-life, representative data from challenging sub-sea environments with low lighting conditions, sand agitation, sea-life and vegetation. The network outputs are configured to perform multi-label image classifications for critical events. The annotation performance varies between 95.1% and 99.7% in terms of accuracy and 90.4% and 99.4% in terms of F1-Score depending on event type. The performance results are on a per-frame basis and corroborate the potential of the algorithm to be the foundation for an intelligent decision support framework that automates the annotation process. The solution can execute annotations in real-time and is significantly more cost-effective than human-only approaches.Entities:
Keywords: deep learning; multi-label image classification; sub-sea pipeline survey; transfer learning; visual inspection
Year: 2020 PMID: 31991872 PMCID: PMC7038356 DOI: 10.3390/s20030674
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Examples of events in subsea pipeline surveys with varying scene conditions; from left to right: burial, exposure, anode, field joint, free span.
Figure 2Label distribution of a total 23,570 frames of the complete dataset.
Figure 3ResNet-50 architecture with modified head.
Figure 4Model training and evaluation process.
Figure 5Ground truth label, image, heatmap and predicted confidence scores for the five different event types.
Figure 6Steps for evaluating model’s performance: (1) validation set, (2) feature extraction, (3) classifier, (4) precision–recall curves for optimal thresholds selection, (5) applying optimal thresholds, (6) comparison with ground truth.
Figure 7Precision–recall curves for all labels. The inset shows a zoomed version of the top right corner.
Optimum label-based thresholds for the validation set.
| Event | Anode | Burial | Exposure | Field Joint | Free Span |
|---|---|---|---|---|---|
|
| 0.357 | 0.367 | 0.632 | 0.542 | 0.430 |
Aggregate performance of the five models, one for each fold.
| Fold # | Exact Match Ratio | Precision | Recall | F1-Score |
|---|---|---|---|---|
| 1 | 0.907 | 0.958 | 0.961 | 0.960 |
| 2 | 0.890 | 0.949 | 0.956 | 0.953 |
| 3 | 0.920 | 0.972 | 0.961 | 0.967 |
| 4 | 0.914 | 0.962 | 0.967 | 0.964 |
| 5 | 0.899 | 0.954 | 0.958 | 0.956 |
Metrics with optimal thresholds on the validation set.
| Threshold | Accuracy | Recall | Precision | F1-Score | |||||
|---|---|---|---|---|---|---|---|---|---|
| Event | Average | Std | Average | Std | Average | Std | Average | Std | |
|
| 0.357 | 0.981 | 0.006 | 0.910 | 0.028 | 0.912 | 0.046 | 0.911 | 0.028 |
|
| 0.367 | 0.978 | 0.001 | 0.959 | 0.011 | 0.953 | 0.013 | 0.956 | 0.004 |
|
| 0.632 | 0.978 | 0.001 | 0.984 | 0.004 | 0.986 | 0.003 | 0.985 | 0.001 |
|
| 0.542 | 0.942 | 0.008 | 0.893 | 0.020 | 0.885 | 0.024 | 0.889 | 0.015 |
|
| 0.430 | 0.995 | 0.002 | 0.988 | 0.002 | 0.988 | 0.013 | 0.988 | 0.007 |
|
| 0.906 | 0.011 | 0.961 | 0.004 | 0.959 | 0.008 | 0.960 | 0.005 | |
Figure 8Confusion matrices on the test set for each class; anode, burial, exposure, field joint and free span.
Test set performance of individual labels and aggregate.
| Event | Threshold | Accuracy | Precision | Recall | F1-Score |
|---|---|---|---|---|---|
| Anode | 0.357 | 0.986 | 0.952 | 0.912 | 0.931 |
| Burial | 0.367 | 0.980 | 0.955 | 0.966 | 0.961 |
| Exposure | 0.632 | 0.980 | 0.988 | 0.984 | 0.986 |
| Field Joint | 0.542 | 0.951 | 0.928 | 0.882 | 0.904 |
| Free Span | 0.430 | 0.997 | 0.997 | 0.990 | 0.994 |
|
|
|
|
|
|
Test set performance of different ResNet model sizes.
| Network | # Parameters | Inference Time (ms) | Exact Match Ratio | Precision | Recall | F1-Score |
|---|---|---|---|---|---|---|
|
| 11,706,949 | 17.7 | 0.872 | 0.945 | 0.947 | 0.946 |
|
| 21,815,109 | 20.8 | 0.903 | 0.953 | 0.966 | 0.960 |
|
| 25,617,477 | 23.6 | 0.919 | 0.972 | 0.960 | 0.966 |
|
| 44,609,605 | 31.2 | 0.916 | 0.956 | 0.973 | 0.965 |
|
| 60,253,253 | 39.1 | 0.833 | 0.931 | 0.927 | 0.929 |