| Literature DB >> 35243030 |
Sebastiaan R S Arends1, Mark H F Savenije1,2, Wietse S C Eppinga1, Joanne M van der Velden1, Cornelis A T van den Berg1,2, Joost J C Verhoeff1.
Abstract
BACKGROUND ANDEntities:
Keywords: Artificial intelligence; Auto-segmentation; Deep learning; Spinal metastases
Year: 2022 PMID: 35243030 PMCID: PMC8857663 DOI: 10.1016/j.phro.2022.02.003
Source DB: PubMed Journal: Phys Imaging Radiat Oncol ISSN: 2405-6316
Fig. 1Overview of networks and approaches. Different colors represent different vertebral labels. Left: sagittal CT image of the thoracic spine. Center: black and white projection of output from the binary network and color output from the labeling network where every vertebra has a distinct color. Right top: sequential approach output from binary and labeling networks; and right bottom: combined approach output.
Fig. 2Examples of automatic delineations (using the combined approach) before and after post-processing, projected on the corresponding sagittal CT image in bone setting (W = 2500, L = 1000). Different colors represent different labels. Top row: mixed and incorrect labels in a part of the spine are corrected by post-processing. Bottom row: due to incorrect segmentation of one vertebra (arrows), all vertebrae below are labeled incorrectly even though they were largely correct before.
Quantitative assessment: median (inter-quartile range) Dice similarity coefficient (DSC), Hausdorff distance (HD) and proportion (95% CI) of correctly labeled vertebrae. The sequential approach outperformed the combined approach during the external validation.
| | |||
| Thoracic vertebrae | 96.1 (95.0–96.9) | 96.2 (95.0–97.0) | 0.003 |
| Lumbar vertebrae | 97.6 (97.3–97.9) | 97.6 (97.1–97.9) | 0.004 |
| All vertebrae | 96.7 (95.5–97.4) | 96.7 (95.4–97.4) | 0.13 |
| | |||
| Thoracic vertebrae | 3.7 (2.8–5.4) | 4.1 (2.8–5.8) | 0.45 |
| Lumbar vertebrae | 3.2 (2.4–4.0) | 3.0 (2.2–4.2) | 0.72 |
| All vertebrae | 3.6 (2.8–5.1) | 3.6 (2.4–5.7) | 0.66 |
| | |||
| All vertebrae | 90.7 (88.3–93.1) | 91.6 (89.3–93.8) | |
| | |||
| Thoracic vertebrae | 93.4 (90.7–94.8) | 93.9 (88.9–95.1) | 0.09 |
| Lumbar vertebrae | 96.0 (95.0–96.4) | 95.4 (92.8–96.1) | <0.001 |
| All vertebrae | 94.5 (91.8–95.8) | 94.4 (91.4–95.5) | <0.001 |
| | |||
| Thoracic vertebrae | 4.6 (3.6–5.9) | 7.1 (3.6–21.3) | <0.001 |
| Lumbar vertebrae | 4.0 (3.2–6.1) | 7.2 (4.0–12.4) | <0.001 |
| All vertebrae | 4.5 (3.4–6.0) | 7.1 (3.7–15.1) | <0.001 |
| | |||
| All vertebrae | 79.6 (74.0–85.2) | 55.7 (48.9–62.6) | |
Fig. 3Quantitative assessment: boxplots of DSC (A) and HD (B) values for both the internal (int.) and external (ext.) validation. Both approaches performed similarly during the internal validation, but the sequential approach outperformed the combined approach during the external validation.
Fig. 4Subjective assessment of contours by radiation oncologists. DL contours were more often rated as having obvious errors (1% vs. 11%, p = 0.004), whereas human-made contours were more frequently considered precise (61% vs. 52%, p = 0.40). In total, 88% of human-made contours were deemed clinically acceptable, compared to 77% of automatic contours.