| Literature DB >> 35408337 |
Yanfen Li1, Hanxiang Wang1, L Minh Dang2, Hyoung-Kyu Song2, Hyeonjoon Moon1.
Abstract
Due to the advantages of economics, safety, and efficiency, vision-based analysis techniques have recently gained conspicuous advancements, enabling them to be extensively applied for autonomous constructions. Although numerous studies regarding the defect inspection and condition assessment in underground sewer pipelines have presently emerged, we still lack a thorough and comprehensive survey of the latest developments. This survey presents a systematical taxonomy of diverse sewer inspection algorithms, which are sorted into three categories that include defect classification, defect detection, and defect segmentation. After reviewing the related sewer defect inspection studies for the past 22 years, the main research trends are organized and discussed in detail according to the proposed technical taxonomy. In addition, different datasets and the evaluation metrics used in the cited literature are described and explained. Furthermore, the performances of the state-of-the-art methods are reported from the aspects of processing accuracy and speed.Entities:
Keywords: computer vision; condition assessment; defect inspection; sewer pipes; survey
Mesh:
Year: 2022 PMID: 35408337 PMCID: PMC9002734 DOI: 10.3390/s22072722
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1There are five stages in the defect inspection framework, which include (a) the data acquisition stage based on various sensors (CCTV, sonar, or scanner), (b) the data processing stage for the collected data, (c) the defect inspection stage containing different algorithms (defect classification, detection, and segmentation), (d) the risk assessment for detected defects using image post-processing, and (e) the final report generation stage for the condition evaluation.
Major contributions of the previous review papers on defect inspection and condition assessment. ‘√’ indicates the research areas (defect inspection or condition assessment) are involved. ‘×’ means the research areas (defect inspection or condition assessment) are not involved.
| ID | Ref. | Time | Defect Inspection | Condition Assessment | Contributions |
|---|---|---|---|---|---|
| 1 | [ | 2019 | √ | √ |
Analyze the status of practical defect detection and condition assessment technologies. Compare the benefits and drawbacks of the reviewed work. |
| 2 | [ | 2020 | √ | × |
Introduce defect inspection methods that are suitable for different materials. Provide a taxonomy of various defects. List the state-of-the-art (SOTA) methods for the classification and detection. |
| 3 | [ | 2020 | √ | × |
Create a brief overview of the defect inspection algorithms, datasets, and evaluation metrics. Indicate three recommendations for the future research. |
| 4 | [ | 2020 | × | √ |
Investigate different models for the condition assessment. Analyze the influence factors of the reviewed models on the sewer conditions. |
| 5 | [ | 2021 | √ | √ |
Present a review for main applications, advantages, and possible research areas. |
Figure 2Number of journal publications investigated in different databases.
Figure 3Number of publications investigated in different time periods (from 2000 to 2022).
Figure 4The classification map of the existing algorithms for each category. The dotted boxes represent the main stages of the algorithms.
Academic studies in vision-based defect classification algorithms.
| Time | Methodology | Advantage | Disadvantage | Ref. |
|---|---|---|---|---|
| 2000 | Back-propagation algorithm | Perform well for classification | Slow learning speed | [ |
| 2002 | Neuro-fuzzy algorithm | Good classification efficiency | Weak feature extraction scheme | [ |
| 2006 | Neuro-fuzzy classifier | ● Combines neural network and fuzzy logic concepts | Not an end-to-end model | [ |
| 2009 | Rule-based classifier | Recognize defects under the realistic sewer condition | No real-time recognition | [ |
| 2009 | Rule-based classifier | Addresses realistic defect detection and recognition | Unsatisfactory classification result | [ |
| 2009 | Radial basis network (RBN) | Overall classification accuracy is high | Heavily relies on the pre-engineered results | [ |
| 2012 | Self-organizing map (SOM) | Suitable for large-scale real applications | High computation complexities | [ |
| 2013 | Ensemble classifiers | ● High practicability | Feature extraction and classification are separately implemented | [ |
| 2016 | Random forests | Dramatically reduces the processing time | Processing speed can be improved | [ |
| 2017 | Random forest classifier | Automated fault classification | Poor performance | [ |
| 2017 | Hidden Markov model (HMM) | ● Efficient for numerous patterns of defects | Low classification accuracy | [ |
| 2018 | One-class SVM (OCSVM) | Available for both still images and video sequences | Cannot achieve a standard performance | [ |
| 2018 | Multi-class random forest | Poor classification accuracy | Real-time prediction | [ |
| 2018 | Multiple binary CNNs | ● Good generalization capability | ● Do not support sub-defects classification | [ |
| 2018 | CNN | ● High detection accuracy | Poor performance for the unnoticeable defects | [ |
| 2018 | HMM and CNN | Automatic defect detection and classification in videos | Poor performance | [ |
| 2019 | Single CNN | ● Outperforms the SOTA | Weak performance for fully automatic classification | [ |
| 2019 | Two-level hierarchical CNNs | Can identify the sewer images into different classes | Cannot classify multiple defects in the same image simultaneously | [ |
| 2019 | Deep CNN | ● Classifies defects at different levels | There exists a extremely imbalanced data problem (IDP) | [ |
| 2019 | CNN | Accurate recognition and localization for each defect | Classifies only one defect with the highest probability in an image | [ |
| 2019 | SVM | Reveals the relationship between training data and accuracy | Requires various steps for feature extraction | [ |
| 2020 | SVM | ● Classifies cracks at a sub-category level | Limited to only three crack patterns | [ |
| 2020 | CNN | ● Image pre-processing used for noisy removal and image enhancement | Can classify one defect only | [ |
| 2020 | CNN | Shows great ability for defect classification under various conditions | Limited to recognize the tiny and narrow cracks | [ |
| 2021 | CNN | Is robust against the IDP and noisy factors in sewer images | No multi-label classification | [ |
| 2021 | CNN | Covers defect classification, detection, and segmentation | Weak classification results | [ |
Academic studies in vision-based defect detection algorithms.
| Time | Methodology | Advantage | Disadvantage | Ref. |
|---|---|---|---|---|
| 2004 | Genetic algorithm (GA) and CNN | High average detection rate | Can only detect one type of defect | [ |
| 2014 | Histograms of oriented gradients (HOG) and SVM | Viable and robust algorithm | Complicated image processing steps before detecting defective regions | [ |
| 2018 | Faster R-CNN | ● Explores the influences of several factors for the model performance | Limited to the still images | [ |
| 2018 | Faster R-CNN | Addresses similar object detection problems in industry | Long training time and slow detection speed | [ |
| 2018 | Rule-based detection algorithm | ● Based on image processing techniques | Low detection performance | [ |
| 2019 | YOLO | End-to-end detection workflow | Cannot detect defect at the sub-classes | [ |
| 2019 | YOLOv3 | ● High detection rate | Weak function of output frames | [ |
| 2019 | SSD, YOLOv3, and Faster R-CNN | Automatic detection for the operational defects | Cannot detect the structural defects | [ |
| 2019 | Rule-based detection algorithm | Performs well on the low-resolution images | Requires multiple digital image processing steps | [ |
| 2019 | Kernel-based detector | Promising and reliable results for anomaly detection | Cannot get the true position inside pipelines | [ |
| 2019 | CNN and YOLO | Obtained a considerable reduction in processing speed | Can detect only one type of structural defect | [ |
| 2020 | Faster R-CNN | Can assess the defect severity as well as the pipe condition | Cannot run in real time | [ |
| 2020 | Faster R-CNN | ● Can obtain the number of defects | Requires training two models separately, not an end-to-end framework | [ |
| 2020 | SSD, YOLOv3, and Faster R-CNN | Automated defect detection | Structural defect detection and severity classification are not available | [ |
| 2021 | YOLOv3 | ● Covers defect detection, video interpretation, and text recognition | The ground truths (GTs) are not convincing | [ |
| 2021 | CNN and non-overlapping windows | Outperformed existing models in terms of detection accuracy | Deeper CNN model with better performance requires longer inference time | [ |
| 2021 | Strengthened region proposal network (SRPN) | ● Effectively locate defects | ● Cannot be applied for online processing | [ |
| 2021 | YOLOv2 | Covers defect classification, detection, and segmentation | Weak detection results | [ |
| 2022 | Transformer-based defect detection (DefectTR) | ● Does not require prior knowledge | The robustness and efficiency can be improved for real-world applications | [ |
Academic studies in vision-based defect segmentation algorithms.
| Time | Methodology | Advantage | Disadvantage | Ref. |
|---|---|---|---|---|
| 2005 | Mathematical morphology-based Segmentation | ● Automated segmentation based on geometry image modeling | ● Can only segment cracks | [ |
| 2014 | Mathematical morphology-based Segmentation | Requires less data and computing resources to achieve a decent performance | ● Challenging to detect cracks | [ |
| 2019 | DL-based semantic segmentation (DilaSeg-CRF) | ● End-to-end trainable model | Long training time | [ |
| 2020 | DilaSeg-CRF | ● Promising segmentation accuracy | Complicated workflow | [ |
| 2020 | DL-based semantic segmentation | ● Enhances the feature extraction capability | Still exists negative segmentation results | [ |
| 2021 | Feature pyramid networks (FPN) and CNN | Covers defect classification, detection, and segmentation | Weak segmentation results | [ |
| 2022 | DL-based defect segmentation (Pipe-SOLO) | ● Can segment defect at the instance level | Only suitable for still sewer images | [ |
Figure 5Proportions of the investigated studies using different inspection algorithms.
Figure 6Optimal separation hyperplane.
Figure 7A fine-tuned network architecture used for defect classification.
Figure 8The YOLOv3 architecture with the 3-scale prediction mechanism.
Figure 9The model architecture of the SSD model.
Figure 10An architecture of the faster R-CNN developed for defect detection. ‘CNN’ refers to convolutional neural network. ‘ROI’ means region of interest. ‘FC’ is fully connected layer.
Figure 11An architecture of PipeUNet proposed for semantic segmentation.
The detailed information of the latest robots for sewer inspection.
| Name | Company | Pipe Diameter | Camera Feature | Country | Strong Point |
|---|---|---|---|---|---|
| CAM160 ( | Sewer Robotics | 200–500 mm | NA | USA | ● Auto horizon adjustment |
| LETS 6.0 ( | ARIES INDUSTRIES | 150 mm or larger | Self-leveling lateral camera or a Pan and tilt camera | USA | ● Slim tractor profile |
| wolverine® 2.02 | ARIES INDUSTRIES | 150–450 mm | NA | USA | ● Powerful crawler to maneuver obstacles |
| X5-HS ( | EASY-SIGHT | 300–3000 mm | ≥2 million pixels | China | ● High-definition |
| Robocam 6 ( | TAP Electronics | 600 mm or more | Sony 130-megapixel Exmor 1/3-inch CMOS | Korea | ● High-resolution |
| RoboCam Innovation4 | TAP Electronics | 600 mm or more | Sony 130-megapixel Exmor 1/3-inch CMOS | Korea | ● Best digital record performance |
| Robocam 30004 | TAP Electronics’ Japanese subsidiary | 250–3000 mm | Sony 1.3-megapixel Exmor CMOS color | Japan | ● Can be utilized in huge pipelines |
Figure 12Representative inspection robots for data acquisition. (a) LETS 6.0, (b) Robocam 6, (c) X5-HS, and (d) Robocam 3000.
Figure 13Sample images from the Sewer-ML dataset that has a wide diversity of materials and shapes.
Research datasets for sewer defects in recent studies.
| ID | Defect Type | Image Resolution | Equipment | Number of Images | Country | Ref. |
|---|---|---|---|---|---|---|
| 1 | Broken, crack, deposit, fracture, hole, root, tap | NA | NA | 4056 | Canada | [ |
| 2 | Connection, crack, debris, deposit, infiltration, material change, normal, root | 1440 × 720–320 × 256 | RedZone® | 12,000 | USA | [ |
| 3 | Attached deposit, defective connection, displaced joint, fissure, infiltration, ingress, intruding connection, porous, root, sealing, settled deposit, surface | 1040 × 1040 | Front-facing and back-facing camera with a 185∘ wide lens | 2,202,582 | The Netherlands | [ |
| 4 | Dataset 1: defective, normal | NA | NA | 40,000 | China | [ |
| Dataset 2: barrier, deposit, disjunction, fracture, stagger, water | 15,000 | |||||
| 5 | Broken, deformation, deposit, other, joint offset, normal, obstacle, water | 1435 × 1054–296 × 166 | NA | 18,333 | China | [ |
| 6 | Attached deposits, collapse, deformation, displaced joint, infiltration, joint damage, settled deposit | NA | NA | 1045 | China | [ |
| 7 | Circumferential crack, longitudinal crack, multiple crack | 320 × 240 | NA | 335 | USA | [ |
| 8 | Debris, joint faulty, joint open, longitudinal, protruding, surface | NA | Robo Cam 6 with a 1/3-in. SONY Exmor CMOS camera | 48,274 | South Korea | [ |
| 9 | Broken, crack, debris, joint faulty, joint open, normal, protruding, surface | 1280 × 720 | Robo Cam 6 with a megapixel Exmor CMOS sensor | 115,170 | South Korea | [ |
| 10 | Crack, deposit, else, infiltration, joint, root, surface | NA | Remote cameras | 2424 | UK | [ |
| 11 | Broken, crack, deposit, fracture, hole, root, tap | NA | NA | 1451 | Canada | [ |
| 12 | Crack, deposit, infiltration, root | 1440 × 720–320 × 256 | RedZone® Solo CCTV crawler | 3000 | USA | [ |
| 13 | Connection, fracture, root | 1507 × 720–720 × 576 | Front facing CCTV cameras | 3600 | USA | [ |
| 14 | Crack, deposit, root | 928 × 576–352 × 256 | NA | 3000 | USA | [ |
| 15 | Crack, deposit, root | 512 × 256 | NA | 1880 | USA | [ |
| 16 | Crack, infiltration, joint, protruding | 1073 × 749–296 × 237 | NA | 1106 | China | [ |
| 17 | Crack, non-crack | 64 × 64 | NA | 40,810 | Australia | [ |
| 18 | Crack, normal, spalling | 4000 × 46,000–3168 × 4752 | Canon EOS. Tripods and stabilizers | 294 | China | [ |
| 19 | Collapse, crack, root | NA | SSET system | 239 | USA | [ |
| 20 | Clean pipe, collapsed pipe, eroded joint, eroded lateral, misaligned joint, perfect joint, perfect lateral | NA | SSET system | 500 | USA | [ |
| 21 | Cracks, joint, reduction, spalling | 512 × 512 | CCTV or Aqua Zoom camera | 1096 | Canada | [ |
| 22 | Defective, normal | NA | CCTV (Fisheye) | 192 | USA | [ |
| 23 | Deposits, normal, root | 1507 × 720–720 × 576 | Front-facing CCTV cameras | 3800 | USA | [ |
| 24 | Crack, non-crack | 240 × 320 | CCTV | 200 | South Korea | [ |
| 25 | Faulty, normal | NA | CCTV | 8000 | UK | [ |
| 26 | Blur, deposition, intrusion, obstacle | NA | CCTV | 12,000 | NA | [ |
| 27 | Crack, deposit, displaced joint, ovality | NA | CCTV (Fisheye) | 32 | Qatar | [ |
| 29 | Crack, non-crack | 320 × 240–20 × 20 | CCTV | 100 | NA | [ |
| 30 | Barrier, deposition, distortion, fraction, inserted | 600 × 480 | CCTV and quick-view (QV) cameras | 10,000 | China | [ |
| 31 | Fracture | NA | CCTV | 2100 | USA | [ |
| 32 | Broken, crack, fracture, joint open | NA | CCTV | 291 | China | [ |
Overview of the evaluation metrics in the recent studies.
| Metric | Description | Ref. |
|---|---|---|
| Precision | The proportion of positive samples in all positive prediction samples | [ |
| Recall | The proportion of positive prediction samples in all positive samples | [ |
| Accuracy | The proportion of correct prediction in all prediction samples | [ |
| F1-score | Harmonic mean of precision and recall | [ |
| FAR | False alarm rate in all prediction samples | [ |
| True accuracy | The proportion of all predictions excluding the missed defective images among the entire actual images | [ |
| AUROC | Area under the receiver operator characteristic (ROC) curve | [ |
| AUPR | Area under the precision-recall curve | [ |
| mAP | mAP first calculates the average precision values for different recall values for one class, and then takes the average of all classes | [ |
| Detection rate | The ratio of the number of the detected defects to total number of defects | [ |
| Error rate | The ratio of the number of mistakenly detected defects to the number of non-defects | [ |
| PA | Pixel accuracy calculating the overall accuracy of all pixels in the image | [ |
| mPA | The average of pixel accuracy for all categories | [ |
| mIoU | The ratio of intersection and union between predictions and GTs | [ |
| fwIoU | Frequency-weighted IoU measuring the mean IoU value weighing the pixel frequency of each class | [ |
Performances of different algorithms in terms of different evaluation metrics.
| ID | Number of Images | Algorithm | Task | Performance | Ref. | |
|---|---|---|---|---|---|---|
| Accuracy (%) | Processing Speed | |||||
| 1 | 3 classes | Multiple binary CNNs | Classification | Accuracy: 86.2 | NA | [ |
| 2 | 12 classes | Single CNN | Classification | AUROC: 87.1 | NA | [ |
| 3 | Dataset 1: 2 classes | Two-level hierarchical CNNs | Classification | Accuracy: 94.5 | 1.109 h for 200 videos | [ |
| Dataset 2: 6 classes | Accuracy: 94.96 | |||||
| 4 | 8 classes | Deep CNN | Classification | Accuracy: 64.8 | NA | [ |
| 5 | 6 classes | CNN | Classification | Accuracy: 96.58 | NA | [ |
| 6 | 8 classes | CNN | Classification | Accuracy: 97.6 | 0.15 s/image | [ |
| 7 | 7 classes | Multi-class random forest | Classification | Accuracy: 71 | 25 FPS | [ |
| 8 | 7 classes | SVM | Classification | Accuracy: 84.1 | NA | [ |
| 9 | 3 classes | SVM | Classification | Recall: 90.3 | 10 FPS | [ |
| 10 | 3 classes | CNN | Classification | Accuracy: 96.7 | 15 min 30 images | [ |
| 11 | 3 classes | RotBoost and statistical feature vector | Classification | Accuracy: 89.96 | 1.5 s/image | [ |
| 12 | 7 classes | Neuro-fuzzy classifier | Classification | Accuracy: 91.36 | NA | [ |
| 13 | 4 classes | Multi-layer perceptions | Classification | Accuracy: 98.2 | NA | [ |
| 14 | 2 classes | Rule-based classifier | Classification | Accuracy: 87 | NA | [ |
| 15 | 2 classes | OCSVM | Classification | Accuracy: 75 | NA | [ |
| 16 | 4 classes | CNN | Classification | Recall: 88 | NA | [ |
| 17 | 2 class | Rule-based classifier | Classification | Accuracy: 84 | NA | [ |
| 18 | 4 classes | RBN | Classification | Accuracy: 95 | NA | [ |
| 19 | 7 classes | YOLOv3 | Detection | mAP: 85.37 | 33 FPS | [ |
| 20 | 4 classes | Faster R-CNN | Detection | mAP: 83 | 9 FPS | [ |
| 21 | 3 classes | Faster R-CNN | Detection | mAP: 77 | 110 ms/image | [ |
| 22 | 3 classes | Faster R-CNN | Detection | Precision: 88.99 | 110 ms/image | [ |
| 23 | 2 classes | CNN | Detection | Accuracy: 96 | 0.2782 s/image | [ |
| 24 | 3 classes | Faster R-CNN | Detection | mAP: 71.8 | 110 ms/image | [ |
| SSD | mAP: 69.5 | 57 ms/image | ||||
| YOLOv3 | mAP: 53 | 33 ms/image | ||||
| 25 | 2 classes | Rule-based detector | Detection | Detection rate: 89.2 | 1 FPS | [ |
| 26 | 2 classes | GA and CNN | Detection | Detection rate: 92.3 | NA | [ |
| 27 | 5 classes | SRPN | Detection | mAP: 50.8 | 153 ms/image | [ |
| 28 | 1 class | CNN and YOLOv3 | Detection | AP: 71 | 65 ms/image | [ |
| 29 | 3 classes | DilaSeg-CRF | Segmentation | PA: 98.69 | 107 ms/image | [ |
| 30 | 4 classes | PipeUNet | Segmentation | mIoU: 76.37 | 32 FPS | [ |
Figure 14Studies on sewer inspections of different classes of defects.