| Literature DB >> 34244879 |
Yang Zhang1, Siwa Chan2,3, Jeon-Hor Chen4,5, Kai-Ting Chang1, Chin-Yao Lin3,6, Huay-Ben Pan7, Wei-Ching Lin8, Tiffany Kwong1,9, Ritesh Parajuli10, Rita S Mehta10, Sou-Hsin Chien3,11, Min-Ying Su12,13.
Abstract
To develop a U-net deep learning method for breast tissue segmentation on fat-sat T1-weighted (T1W) MRI using transfer learning (TL) from a model developed for non-fat-sat images. The training dataset (N = 126) was imaged on a 1.5 T MR scanner, and the independent testing dataset (N = 40) was imaged on a 3 T scanner, both using fat-sat T1W pulse sequence. Pre-contrast images acquired in the dynamic-contrast-enhanced (DCE) MRI sequence were used for analysis. All patients had unilateral cancer, and the segmentation was performed using the contralateral normal breast. The ground truth of breast and fibroglandular tissue (FGT) segmentation was generated using a template-based segmentation method with a clustering algorithm. The deep learning segmentation was performed using U-net models trained with and without TL, by using initial values of trainable parameters taken from the previous model for non-fat-sat images. The ground truth of each case was used to evaluate the segmentation performance of the U-net models by calculating the dice similarity coefficient (DSC) and the overall accuracy based on all pixels. Pearson's correlation was used to evaluate the correlation of breast volume and FGT volume between the U-net prediction output and the ground truth. In the training dataset, the evaluation was performed using tenfold cross-validation, and the mean DSC with and without TL was 0.97 vs. 0.95 for breast and 0.86 vs. 0.80 for FGT. When the final model developed with and without TL from the training dataset was applied to the testing dataset, the mean DSC was 0.89 vs. 0.83 for breast and 0.81 vs. 0.81 for FGT, respectively. Application of TL not only improved the DSC, but also decreased the required training case number. Lastly, there was a high correlation (R2 > 0.90) for both the training and testing datasets between the U-net prediction output and ground truth for breast volume and FGT volume. U-net can be applied to perform breast tissue segmentation on fat-sat images, and TL is an efficient strategy to develop a specific model for each different dataset.Entities:
Keywords: Breast segmentation; Deep learning; Fibroglandular tissue segmentation; Transfer learning; U-net
Mesh:
Year: 2021 PMID: 34244879 PMCID: PMC8455741 DOI: 10.1007/s10278-021-00472-z
Source DB: PubMed Journal: J Digit Imaging ISSN: 0897-1889 Impact factor: 4.056
Fig. 1Architecture of the fully convolutional residual neural network (FC-RNN), or U-net. The U-net consists of convolution and max-pooling layers at the descending phase (the initial part of the U), the down-sampling stage. At the ascending part of the network, up-sampling operations are performed, which are also implemented by convolutions, where kernel weights are learned during training. The arrows between the two parts show the incorporation of the information
available at the down-sampling steps into the up-sampling operations. The input of the network is the normalized image, and the output is the probability map of the segmentation result
Fig. 2A–D Four representative cases of different breast size and parenchymal pattern showing accurate FGT segmentation using U-net compared to the ground truth. Left column: original image; central column: ground truth of breast and FGT segmentation; right column: segmentation results using U-net. Lower two panels (C and D) show two cases with susceptibility artifacts. Despite of the artifact of bright signal intensity (arrows) similar to FGT, U-net can still recognize and exclude it
Segmentation dice similarity coefficient (DSC) and pixel-based accuracy in training and testing datasets without and with transfer learning
| Dataset/methods | Segmentation Region | DSC | Accuracy | ||
|---|---|---|---|---|---|
| Range* | Mean ± stdev | Range* | Mean ± stdev | ||
| Training set (no transfer learning) | Breast | 0.92–0.99 | 0.95 ± 0.03 | 0.93–0.99 | 0.97 ± 0.04 |
| Fibroglandular Tissue | 0.44–0.92 | 0.80 ± 0.11 | 0.51–0.93 | 0.86 ± 0.03 | |
| Training set (w/ transfer learning) | Breast | 0.96–0.99 | 0.97 ± 0.02 | 0.95–0.99 | 0.97 ± 0.01 |
| Fibroglandular Tissue | 0.33–0.96 | 0.86 ± 0.08 | 0.53–0.98 | 0.90 ± 0.05 | |
| Testing set (no transfer learning) | Breast | 0.69–0.98 | 0.83 ± 0.06 | 0.79–0.98 | 0.89 ± 0.03 |
| Fibroglandular Tissue | 0.34–0.95 | 0.81 ± 0.10 | 0.52–0.98 | 0.87 ± 0.07 | |
| Testing set (w/ transfer learning) | Breast | 0.72–0.98 | 0.89 ± 0.06 | 0.82–0.98 | 0.91 ± 0.03 |
| Fibroglandular Tissue | 0.38–0.97 | 0.81 ± 0.08 | 0.48–0.98 | 0.86 ± 0.05 | |
*Range is the value in the 126 patients in the training dataset and 40 patients in the testing dataset
Fig. 3Correlation of breast volume between the ground truth obtained from the template-based segmentation method and the U-net prediction. A Training data breast volumes. B Training data FGT volumes. C Testing data breast volumes. D Testing data FGT volumes. The red line is the trend line, and the dashed black line is the unity line as reference
Fig. 4A–D Four cases of inconsistent FGT segmentation between U-net and the ground truth. Left column: original image; central column: ground truth of breast and FGT segmentation; right column: segmentation results using U-net. A and B cases show that the FGT results from ground truth are over-segmented compared to the original image. The results clearly show the superior accuracy of U-net. C and D cases show that the FGT results of the ground truth are under-segmented compared to the original image. Note the under-segmented FGT in the lower margin (yellow arrows) of the D case. Note also the incomplete suppression of the fat signals (red arrows) which are recognized and excluded by U-net
Fig. 5The plot of DSC in the testing dataset by using the model developed with different number of training cases from 10, 20, … to 126, with and without TL. When the training case number is small, DSC is low. When sufficient number of cases is used for training (> 30 or breast segmentation, and > 80 for FGT segmentation), the achieved DSC with and without TL is comparable, only slightly better with TL for breast segmentation