| Literature DB >> 29340289 |
Ester Bonmati1,2,3, Yipeng Hu1,2,3, Nikhil Sindhwani4, Hans Peter Dietz5, Jan D'hooge4, Dean Barratt1,2,3, Jan Deprest2,4, Tom Vercauteren1,2,3,4.
Abstract
Segmentation of the levator hiatus in ultrasound allows the extraction of biometrics, which are of importance for pelvic floor disorder assessment. We present a fully automatic method using a convolutional neural network (CNN) to outline the levator hiatus in a two-dimensional image extracted from a three-dimensional ultrasound volume. In particular, our method uses a recently developed scaled exponential linear unit (SELU) as a nonlinear self-normalizing activation function, which for the first time has been applied in medical imaging with CNN. SELU has important advantages such as being parameter-free and mini-batch independent, which may help to overcome memory constraints during training. A dataset with 91 images from 35 patients during Valsalva, contraction, and rest, all labeled by three operators, is used for training and evaluation in a leave-one-patient-out cross validation. Results show a median Dice similarity coefficient of 0.90 with an interquartile range of 0.08, with equivalent performance to the three operators (with a Williams' index of 1.03), and outperforming a U-Net architecture without the need for batch normalization. We conclude that the proposed fully automatic method achieved equivalent accuracy in segmenting the pelvic floor levator hiatus compared to a previous semiautomatic approach.Entities:
Keywords: automatic segmentation; convolutional neural network; levator hiatus; self-normalizing neural network; ultrasound
Year: 2018 PMID: 29340289 PMCID: PMC5762003 DOI: 10.1117/1.JMI.5.2.021206
Source DB: PubMed Journal: J Med Imaging (Bellingham) ISSN: 2329-4302
Fig. 1Network architecture, where and correspond to the spatial dimension and to the number of channels. For the U-Net, the SELU unit is replaced by batch normalization and ReLU, and for the U-Net with dilated convolution (U-Net + DC), the last layer is also replaced by a dilated convolution.
Fig. 2(a) SU-Net architecture versus (b) U-Net architecture.
Performance of the SU-Net, SU-Net + dropout, U-Net, U-Net + DC, and ResNet networks by employing a pairwise comparison with the three manual labels available for each ultrasound image. This table also contains results from a previous study (Sindhwani et al.). Results are reported using median (interquartile range).
| Method | Dice | Jaccard | Hausdorff (in mm) | MAD (in mm) | SMAD (in mm) | FPD | FND |
|---|---|---|---|---|---|---|---|
| SU-Net | 0.90 (0.08) | 0.82 (0.12) | 4.21 (3.92) | 1.19 (1.15) | 1.16 (1.02) | 0.07 (0.13) | 0.09 (0.16) |
| SU-Net + dropout | 0.90 (0.08) | 0.81 (0.13) | 3.90 (3.83) | 1.21 (1.16) | 1.23 (1.09) | 0.07 (0.13) | 0.09 (0.16) |
| U-Net | 0.89 (0.11) | 0.80 (0.18) | 4.49 (5.67) | 1.31 (1.42) | 1.34 (1.41) | 0.07 (0.16) | 0.08 (0.16) |
| U-Net + DC | 0.90 (0.08) | 0.82 (0.13) | 3.97 (3.87) | 1.18 (3.86) | 1.17 (1.23) | 0.05 (0.13) | 0.11 (0.15) |
| ResNet | 0.91 (0.08) | 0.83 (0.14) | 3.59 (4.22) | 1.13 (1.14) | 1.10 (1.07) | 0.06 (0.14) | 0.07 (0.13) |
| Sindhwani et al. | 0.92 (0.05) | 0.85 (0.09) | 5.73 (3.90) | 2.10 (1.54) | — | — | — |
Differences between the manual labels from the three operators (i.e., IOD). Results are reported using median (interquartile range).
| Dice | Jaccard | Hausdorff (in mm) | MAD (in mm) | SMAD (in mm) | FPD | FND |
|---|---|---|---|---|---|---|
| 0.92 (0.06) | 0.85 (0.10) | 3.05 (2.33) | 1.01 (0.85) | 1.01 (0.81) | 0.03 (0.08) | 0.08 (0.15) |
WIs (95% CI) for the SU-Net, SU-Net + dropout, U-Net, U-Net + DC, and ResNet architectures for each evaluation metric. A CI containing the value 1.0 indicates a good agreement between the automatic method and the three operators.
| Method | WI Dice | WI Jaccard | WI Hausdorff (in mm) | WI MAD (in mm) | WI SMAD (in mm) | WI FPD | WI FND |
|---|---|---|---|---|---|---|---|
| SU-Net | 1.032 (1.03, 1.03) | 1.052 (1.05, 1.06) | 0.677 (0.67, 0.69) | 0.738 (0.73, 0.75) | 0.776 (0.77, 0.79) | 0.425 (0.40, 0.45) | 0.588 (0.57, 0.61) |
| SU-Net + dropout | 1.032 (1.03, 1.03) | 1.051 (1.05, 1.05) | 0.701 (0.69, 0.71) | 0.751 (0.74, 0.76) | 0.784 (0.77, 0.80) | 0.420 (0.40, 0.44) | 0.591 (0.57, 0.62) |
| U-Net | 1.085 (1.08, 1.09) | 1.111 (1.10, 1.12) | 0.530 (0.52, 0.54) | 0.577 (0.56, 0.59) | 0.538 (0.52, 0.56) | 0.281 (0.26, 0.30) | 0.439 (0.42, 0.46) |
| U-Net + DC | 1.033 (1.03, 1.04) | 1.053 (1.05, 1.06) | 0.712 (0.70, 0.72) | 0.723 (0.71, 0.74) | 0.756 (0.74, 0.77) | 0.395 (0.37, 0.42) | 0.706 (0.69, 0.72) |
| ResNet | 1.037 (1.03, 1.04) | 1.061 (1.06, 1.07) | 0.717 (0.71, 0.73) | 0.726 (0.71, 0.74) | 0.731 (0.72, 0.74) | 0.533 (0.50, 0.57) | 0.52 (0.5, 0.54) |
COD and IOD using SU-Net with the corresponding WIs and the 95% CI. Results are reported using mean ().
| Stage | Contraction | Valsalva | Rest |
|---|---|---|---|
| COD | |||
| IOD | |||
| WI | 0.80 | 0.72 | 0.85 |
| (95% CI) | (0.72, 0.89) | (0.68, 0.76) | (0.80, 0.90) |
Fig. 3Segmentation of the levator hiatus using with the SU-Net architecture (blue) compared with the three manual labels (red) for the following percentiles of the Dice coefficient: (a) 0th, (b) 25th, (c) 50th, (d) 75th, and (e) 100th.
Fig. 4Segmentation examples of the levator hiatus at the three different stages (contraction, Valsalva, and rest) using the proposed method (blue) compared to the outlines provided by the operators (red). Cases were chosen at the 75th percentile of the mean Dice coefficient considering the three operators.
Fig. 5Histogram of the SELU activations at the last block after (a) 500, (b) 1000, (c) 1500, (d) 2000, (e) 2500, and (f) 3000 iterations.
Fig. 6Overlap at different iterations (0 to 3000) for the U-Net (blue) and SU-Net (orange) architectures during testing for the first fold and for the three operators.
Fig. 7Learning curves of the training loss for the U-Net (blue) and SU-Net (orange) architectures averaged for all folds at different iterations (0 to 3000).