| Literature DB >> 32612129 |
Marie Kloenne1,2, Sebastian Niehaus1,3, Leonie Lampe1, Alberto Merola1, Janis Reinelt1, Ingo Roeder3,4, Nico Scherf5,6.
Abstract
Machine learning has considerably improved medical image analysis in the past years. Although data-driven approaches are intrinsically adaptive and thus, generic, they often do not perform the same way on data from different imaging modalities. In particular computed tomography (CT) data poses many challenges to medical image segmentation based on convolutional neural networks (CNNs), mostly due to the broad dynamic range of intensities and the varying number of recorded slices of CT volumes. In this paper, we address these issues with a framework that adds domain-specific data preprocessing and augmentation to state-of-the-art CNN architectures. Our major focus is to stabilise the prediction performance over samples as a mandatory requirement for use in automated and semi-automated workflows in the clinical environment. To validate the architecture-independent effects of our approach we compare a neural architecture based on dilated convolutions for parallel multi-scale processing (a modified Mixed-Scale Dense Network: MS-D Net) to traditional scaling operations (a modified U-Net). Finally, we show that an ensemble model combines the strengths across different individual methods. Our framework is simple to implement into existing deep learning pipelines for CT analysis. It performs well on a range of tasks such as liver and kidney segmentation, without significant differences in prediction performance on strongly differing volume sizes and varying slice thickness. Thus our framework is an essential step towards performing robust segmentation of unknown real-world samples.Entities:
Year: 2020 PMID: 32612129 PMCID: PMC7329868 DOI: 10.1038/s41598-020-67544-y
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Fig. 1Windowing highlights tissues of interests and reduces the complexity of background structures. Three examples for the use of case-oriented windowing for bones (a–c), organs (d–f), and lungs (g–i). We used the organ oriented windowing in this work, while we show the other two examples for comparison. We derived the intensity windows for CNN processing by slightly extending the standard ranges used by radiologists in practice to allow for uncertainties in the exact ranges.
Fig. 2Differences in CT scanning configurations pose challenges for CNN-based segmentation. (a) Varying slice thickness maps the same anatomical region of interest to different numbers of slices. Thicker slices reduce the scan time for larger regions of interest, but 3D details and semantic context can be lost. (b) Volume size varies depending on the chosen region of interest. Normalising to a standardised volume size then requires strong interpolation.
Comparison of the CT-specific image augmentation (CTIA) and the multidimensional image augmentation (MIA).
| Transformation | CTIA | MIA |
|---|---|---|
| Spatial transformation | Random patch extraction | Scaling |
| Slice skipping | Random rotation | |
| Slice interpolation | Without restriction | |
| Random rotation (maximum of 16°) | Image shearing | |
| Cropping | ||
| Intensity transformations | Cluster-wise voxel intensity range shift | Gamma-corrections |
| Contrast | ||
| Image noising with Gaussian noise | Brightness | |
| Image noising | ||
| with Gaussian noise |
Fig. 3Overview of the different segmentation workflows that we considered in our experiments. The arrows (both solid and dashed) indicate different combinations of input dimension, augmentation and CNN architectures. The solid arrows specifically highlight the best combination.
Results for the kidney tumor segmentation: Total Dice scores are reported (mean ± stdv.) for each segmentation class, the different architectures and input dimensionalities (2D and 3D). Each approach is validated with the multidimensional image augmentation (MIA) for Tensorflow and with our CT-specific image augmentation (CTIA).
| Kidney | Tumor | Total | ||
|---|---|---|---|---|
| nnU-Net + MIA | 2D | |||
| nnU-Net + CTIA | 2D | |||
| nnU-Net + MIA | 3D | |||
| nnU-Net + CTIA | 3D | |||
| MS-D Net + MIA | 2D | |||
| MS-D Net + CTIA | 2D | |||
| MS-D Net + MIA | 3D | |||
| MS-D Net + CTIA | 3D | |||
| Stacked CNN + MIA | ||||
| Stacked CNN + CTIA |
Results for liver segmentation: Total Dice score (mean ± stdv.) for the different architectures and input dimensionalities (2D and 3D). We validated each approach with the multidimensional image augmentation (MIA) for Tensorflow and with our CT-specific image augmentation (CTIA).
| Total | ||
|---|---|---|
| nnU-Net + MIA | 2D | |
| nnU-Net + CTIA | 2D | |
| nnU-Net + MIA | 3D | |
| nnU-Net + CTIA | 3D | |
| MS-D Net + MIA | 2D | |
| MS-D Net + CTIA | 2D | |
| MS-D Net + MIA | 3D | |
| MS-D Net + CTIA | 3D | |
| Stacked CNN + MIA | ||
| Stacked CNN + CTIA |
Fig. 4Examples of challenging 2D segmentation cases. Examples are shown for kidney and tumor segmentation (a) and liver segmentation (b). Segmentation errors typically occur more frequently in the first and last slice of the ROI.
Fig. 5The influence of data augmentation on segmentation quality. Typical examples of low-quality segmentation results of a 2D U-Net trained with (a–d) MIA and (e–h) CTIA. The arrows in the magnified results in (c) highlight incorrect tissue classifications obtained by training with MIA. The reference segmentation is depicted in (d) for comparison. In contrast, segmentation errors in the pipeline trained with CTIA are typically limited to the size of the segmented region (g) as compared to the reference shown in (h). Both examples demonstrate cases with an accuracy of less than one standard deviations below the mean for the respective pipeline. The examples were selected randomly from a pool of examples of the same quality. The effects shown occur for 2D and 3D segmentation models alike (see Supplementary Fig. S1 online) for further examples.