| Literature DB >> 31496928 |
Anna Kuzina1, Evgenii Egorov1, Evgeny Burnaev1.
Abstract
Automatic segmentation methods based on deep learning have recently demonstrated state-of-the-art performance, outperforming the ordinary methods. Nevertheless, these methods are inapplicable for small datasets, which are very common in medical problems. To this end, we propose a knowledge transfer method between diseases via the Generative Bayesian Prior network. Our approach is compared to a pre-train approach and random initialization and obtains the best results in terms of Dice Similarity Coefficient metric for the small subsets of the Brain Tumor Segmentation 2018 database (BRATS2018).Entities:
Keywords: 3D CNN; Bayesian neural networks; brain lesion segmentation; brain tumor segmentation; transfer learning; variational autoencoder
Year: 2019 PMID: 31496928 PMCID: PMC6712162 DOI: 10.3389/fnins.2019.00844
Source DB: PubMed Journal: Front Neurosci ISSN: 1662-453X Impact factor: 4.677
Figure 1U-Net architecture used in the experiments contains ResNet-like blocks both in Encoder and Decoder parts with skip-connections.
Algorithm for training model with Deep Weight Prior.
| while not converged |
| Sample minibatch |
| |
| Sample weights with reparametrization: |
| Sample latent variables with reparametrization: |
| Compute stochastic gradients of the objective: |
| |
| Update parameters |
Figure 2Full scheme of the training procedure with Deep Weight Prior.
Figure 3Example of MRI slices and ground truth segmentation from BRATS18 dataset.
Figure 4Example of MRI slices and ground truth segmentation from MS dataset.
Procedure for UNet-PR training on m images
| Train one 3D U-Net model on |
| |
| Split |
| Select |
| Initialize model with the weights trained on |
| Train 3D U-Net on selected images |
| Evaluate model on |
Procedure of UNet-DWP training on m images
| Train 3D U-Net models with different initializations on |
| Collect kernels and split them into seven parts (depending on the input size of the layer) |
| Train 7 VAE, to use them as implicit prior |
| |
| Split |
| Select |
| Train 3D U-Net on selected images with DWP |
| Evaluate model on |
U-Net hyperparameters details.
| Batch-size | 2 |
| Optimizer | Adam |
| Initial learning rate | 10−3 |
| LR scheduler | Reduce learning rate when a loss has stopped improving |
| LR scheduler patience | 10 |
| LR scheduler factor | 0.1 |
| Max epochs | 500 |
| Early stopping criterion | LR == 10−6 |
| Test size | 50 |
| Train sizes | [5, 10, 15, 20] |
Figure 5Examples of trained kernels. (A) Kernels from U-Net, trained on MS dataset, which were further used to train DWP. (B) Samples from trained Deep Weight Prior.
DWP hyperparameters details.
| Batch-size | 20 |
| Optimizer | Adam |
| Initial learning rate | 10−3 |
| LR scheduler | Reduce learning rate when a loss has stopped improving |
| LR scheduler patience | 15 |
| LR scheduler factor | 0.1 |
| Max epochs | 500 |
| Early stopping criterion | LR == 10−6 |
| Latent dimension | 6 |
Mean Dice Similarity Score for the different subsets of BRATS18 dataset.
| 5 | 0.61 (0.02) | 0.58 (0.03) | 0.62 (0.02) | |
| 10 | 0.64 (0.01) | 0.60 (0.03) | 0.66 (0.01) | |
| 15 | 0.67 (0.02) | 0.63 (0.02) | 0.70 (0.02) | |
| 20 | 0.69 (0.01) | 0.65 (0.02) | 0.70 (0.01) |
Best results are in bold.
Mean Intersection over Union for the different subsets of BRATS18 dataset.
| 5 | 0.49 (0.02) | 0.45 (0.03) | 0.50 (0.02) | |
| 10 | 0.52 (0.01) | 0.47 (0.03) | 0.53 (0.01) | |
| 15 | 0.56 (0.02) | 0.50 (0.02) | 0.58 (0.02) | |
| 20 | 0.58 (0.01) | 0.53 (0.02) | 0.60 (0.01) |
Best results are in bold.
Figure 6Segmentation accuracy on BRATS18 dataset for various train sample size, calculated for three different splits.
Figure 7Examples of models' predictions on test samples, compared to ground truth segmentation. (A) Test MRI. (B) Ground truth segmentation. (C) UNet-DWP. (D) UNet-PR. (E) UNet-RI.
Mean Dice Similarity Score for the subsets of Task03_Liver and Task09_Spleen datasets.
| 5 | 0.275 | 0.209 | 0.391 | 0.105 | ||
| 10 | 0.293 | 0.052 | 0.584 | 0.239 | ||
| 15 | 0.306 | 0.243 | 0.556 | 0.302 | ||
| 20 | 0.336 | 0.156 | 0.566 | 0.459 | ||
Best results are in bold.
Mean Dice Similarity Score for the experiments with large available target dataset (MS-BRATS18).
| 100 | 0.76 (0.01) | 0.79 (0.01) | 0.77 (0.01) | 0.77 (0.01) |