| Literature DB >> 34930947 |
Ivo M Baltruschat1, Hanna Ćwieka2, Diana Krüger2, Berit Zeller-Plumhoff2, Frank Schlünzen3, Regine Willumeit-Römer2, Julian Moosmann4, Philipp Heuser3,5.
Abstract
Highly accurate segmentation of large 3D volumes is a demanding task. Challenging applications like the segmentation of synchrotron radiation microtomograms (SRμCT) at high-resolution, which suffer from low contrast, high spatial variability and measurement artifacts, readily exceed the capacities of conventional segmentation methods, including the manual segmentation by human experts. The quantitative characterization of the osseointegration and spatio-temporal biodegradation process of bone implants requires reliable, and very precise segmentation. We investigated the scaling of 2D U-net for high resolution grayscale volumes by three crucial model hyper-parameters (i.e., the model width, depth, and input size). To leverage the 3D information of high-resolution SRμCT, common three axes prediction fusing is extended, investigating the effect of adding more than three axes prediction. In a systematic evaluation we compare the performance of scaling the U-net by intersection over union (IoU) and quantitative measurements of osseointegration and degradation parameters. Overall, we observe that a compound scaling of the U-net and multi-axes prediction fusing with soft voting yields the highest IoU for the class "degradation layer". Finally, the quantitative analysis showed that the parameters calculated with model segmentation deviated less from the high quality results than those obtained by a semi-automatic segmentation method.Entities:
Mesh:
Year: 2021 PMID: 34930947 PMCID: PMC8688506 DOI: 10.1038/s41598-021-03542-y
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1Overview of the segmentation framework for high-resolution synchrotron radiation microtomograms. Top shows the full segmentation pipeline: 1. conversion of 3D tomograms into 2D slices. 2. processing of slicing sets by our model. 3. soft voting is used to fuse the multi-axes prediction into the final segmentation. Top right: 3D rendering of the resulting segmentation (created with 3D Slicer, v4.11, https://www.slicer.org/). Bottom shows our final U-net model architecture and the layer legend (created with Net2Vis[38], https://github.com/viscom-ulm/Net2Vis).
Synchrotron radiation microtomography dataset characteristics.
| 3D Samples | 14 | 3 | |
| 2D Images | 47,600 | – | |
| Mg-5Gd | 4 (29%) | 3 (100%) | |
| Mg-10Gd | 10 (71%) | – | |
| IBL | 9 (64%) | 1 (33%) | |
| I13-2 | 5 (36%) | 2 (66%) | |
For training, each sample is sliced into 2D images based on the three main axes. For testing, the number of 2D images changes based on the multi-axes fusing method.
Figure 2Slicing example (sample 1) for three naive planes (created with 3D Slicer, v4.11, https://www.slicer.org/). From left to right: is show in red, in green, and in yellow. The arrow indicates the slicing direction.
Figure 3Results for scaling of the U-net. Each plot shows the mean IoU (i.e., averaged over the 4-fold cross validation and the classes) vs the floating point operations per second (FLOPs). For compound scaling, we vary the model width for each step of each run. From left to right: model depth d, model width , model input size IS, compound scaling of multiple hyper-parameters.
Mean Intersection over Union (IoU) and the standard error results for different 3D information fusing methods.
| Baseline, | ||||
| Baseline, | ||||
| Baseline, | ||||
| Avr, 3-planes | ||||
| MV, 3-planes | ||||
| Avr, 9-planes | ||||
| MV, 9-planes |
Bold text for each column emphasizes the overall highest mean IoU value. All values are scaled by 100 for convenience.
Figure 4Probability results for processing the 3D volume slice-by-slice and the proposed soft voting fusing method. Here, we show and (in the first and second row, respectively) of the test sample 1. For , each image shows the probability output for “bone” of our best model without conversion to a final segmentation. For , we show the probability output for “residual material”. From left to right: “Baseline, rc”, “Baseline, pc”, “Baseline, pr”, “Avr, 3-planes”, and “Avr, 9-planes”. A high value indicates that this area is most likely “bone” or “degradation layer” for and , respectively.
Figure 5Comparison between different types of segmentation results—high quality (manual), workflow, and machine learning (Avr., 9-planes). We show and (in the first and second row, respectively) of test sample 1. For the segmentation results, the images are colored based on the corresponding label: residual material (RM), degradation layer (DL), bone, and background (BG).
Comparison of the quantified parameters for each type of segmentation.
| DR [mm/year] | 1 | 0.252 (−3%) | 0.243 (−7%) | |
| 2 | 0.205 (−2%) | 0.208 (−) | ||
| 3 | 0.417 (−10%) | 0.436 (−6%) | ||
| BIC [%] | 1 | 62.94 (−) | 70.44 (+12%) | |
| 2 | 81.07 (+1%) | 80.30 (−) | ||
| 3 | 60.13 (+33%) | 51.61 (+14%) | ||
| BV/TV [%] | 1 | 47.88 (−1%) | 48.45 (+1%) | |
| 2 | 55.23 (+1%) | 54.67 (−) | ||
| 3 | 41.25 (+4%) | 41.25 (+4%) |
We consider high quality (HQ) segmentation as the reference (in bold). Percentage values in brackets represent the relative differences between workflow (WF) and machine learning (ML) segmentation compared to the HQ segmentation. (−) means that the difference was less than 1%.