| Literature DB >> 35595948 |
Eva Schnider1, Antal Huck2, Mireille Toranelli3, Georg Rauter2, Magdalena Müller-Gerbl3, Philippe C Cattin2.
Abstract
PURPOSE: Automated distinct bone segmentation has many applications in planning and navigation tasks. 3D U-Nets have previously been used to segment distinct bones in the upper body, but their performance is not yet optimal. Their most substantial source of error lies not in confusing one bone for another, but in confusing background with bone-tissue.Entities:
Keywords: CT; Deep-learning; Distinct bone segmentation; U-Net
Mesh:
Year: 2022 PMID: 35595948 PMCID: PMC9515055 DOI: 10.1007/s11548-022-02650-y
Source DB: PubMed Journal: Int J Comput Assist Radiol Surg ISSN: 1861-6410 Impact factor: 3.421
Fig. 1Volume rendering of one of our upper-body CT scans (left), and the result of our automated segmentation using BEM-inference and label-correction (right)
Fig. 2Results on the synthetic dataset using the baseline 3D U-Net, and Dual D with our proposed BEM-inference. Both false positives (around the elbows), and false negatives (head) are reduced using our approach
Network architectures comparison for the upper-body CT dataset
| Model | Trainable parameters (#) | Training time | Inference time |
|---|---|---|---|
| Baseline 3D U-Net | 0.84 | 219 | |
| Dual A | 1.08 | 212 | |
| Dual B | 1.08 | 271 | |
| Dual C | 1.15 | 243 | |
| Dual D | 1.20 | 321 |
Average time per training iteration on a voxel patch.
Inference time for an average scan ( voxels) , including data I/O time
Fig. 3Schematic of the four network architectures with dual segmentation heads. They are all based on a 3D U-Net architectures with variances of how the binary segmentation head is appended. See also “Dual segmentation head architecture” Section
Fig. 4Schematic of the BEM-inference process. The background class is denoted in gray, the two distinct foreground classes in blue and pink, respectively
Fig. 5Label confusion matrices (row-normalized) for the baseline 3D U-Net and Dual D, including BEM-inference and post-processing. With our approach, less labels are erroneously classified as background (first column)
Upper-body CT dataset: Results in DSC, comparing the segmentation performance when using baseline inference, against our BEM-inference, with and without label correction
| Baseline | + Label correction | + BEM-inference | + Both | |
|---|---|---|---|---|
| Baseline 3D U-Net | ||||
| Two-stage: pred. bin. | ” | ” | ||
| Two-stage: gt bin. | ” | ” | ||
| Dual A | ||||
| Dual B | ||||
| Dual C | ||||
| Dual D |
The comparison is given for the two-stage models and the different flavors of dual-segmentation heads models. For a description of the metrics, see “Evaluation metrics” Section
Fig. 6Segmentation results and typical errors obtained with the baseline U-Net model and our Dual D model with BEM-inference and post-processing. Using the baseline model, ribs are often not segmented as one, but are assigned multiple labels (I). The post-processing remedies this issue visibly. Other frequent errors occur around the border of vertebrae, especially in the presence of calcifications (II). Within big bones such as hips and femurs, we observe holes and islands where the left/right part of the label has been mixed up (III)
Comparison to other published work on distinct bone segmentation
| Ours (median) | [ | [ | |
|---|---|---|---|
| L3 | 0.85 | 0.85 | 0.91 |
| Sacrum | 0.90 | 0.88 | |
| Clavicula | 0.92 | 0.57 | |
| Hamate | 0.86 | ||
| Inference time per scan (min) | |||
| Scans in dataset (#) | 11 | 100 | 19 |
| Classes (#) | 126 | 49 | 62 |
Results in DSC
Synthetic dataset: Results in DSC, comparing the segmentation performance when using baseline inference, against our BEM-inference, with and without label correction
| Model | Baseline | + BEM-inference |
|---|---|---|
| Two-stage: gt binary seg. | ||
| Dual A: parallel losses | ||
| Dual B: parallel final layers | ||
| Dual C: sequential heads | ||
| Dual D: separate decoders |
The comparison is given for the two-stage models and the different flavors of dual-segmentation heads models. For a description of the metrics, see “Evaluation metrics” Section