| Literature DB >> 30510671 |
Abstract
Using deep neural networks for segmenting an MRI image of heterogeneously distributed pixels into a specific class assigning a label to each pixel is the concept of the proposed approach. This approach facilitates the application of the segmentation process on a preprocessed MRI image, with a trained network to be utilized for other test images. As labels are considered expensive assets in supervised training, fewer training images and training labels are used to obtain optimal accuracy. To validate the performance of the proposed approach, an experiment is conducted on other test images (available in the same database) that are not part of the training; the obtained result is of good visual quality in terms of segmentation and quite similar to the ground truth image. The average computed Dice similarity index for the test images is approximately 0.8, whereas the Jaccard similarity measure is approximately 0.6, which is better compared to other methods. This implies that the proposed method can be used to obtain reference images almost similar to the segmented ground truth images.Entities:
Mesh:
Year: 2018 PMID: 30510671 PMCID: PMC6230419 DOI: 10.1155/2018/3640705
Source DB: PubMed Journal: J Healthc Eng ISSN: 2040-2295 Impact factor: 2.682
Figure 1SegNet architecture pictorial representation as presented by Vijay Badrinarayanan et al. [5].
Figure 2Schematic representation of 31 layers and 34 connections used in proposed CNN Network.
Simplified SegNet network.
| S. No. | Layer name | Type | Description |
|---|---|---|---|
| 1 | “Image input” | Image | 208 × 1761 images with “zero center” normalization |
| 2 | “encoder1_conv1” | Convolution | 64 3 × 3 × 1 convolutions with stride [1 1] and padding [1 1 1 1] |
| 3 | “encoder1_bn_1” | Batch normalization | Batch normalization with 64 channels |
| 4 | “encoder1_relu_1” | ReLU | ReLU |
| 5 | “encoder1_conv2” | Convolution | 64 3 × 3 × 64 convolutions with stride [1 1] and padding [1 1 1 1] |
| 6 | “encoder1_bn_2” | Batch normalization | Batch normalization with 64 channels |
| 7 | “encoder1_relu_2” | ReLU | ReLU |
| 8 | “encoder1_maxpool” | Max pooling | 2 × 2 max pooling with stride [2 2] and padding [0 0 0 0] |
| 9 | “encoder2_conv1” | Convolution | 64 3 × 3 × 64 convolutions with stride [1 1] and padding [1 1 1 1] |
| 10 | “encoder2_bn_1” | Batch normalization | Batch normalization with 64 channels |
| 11 | “encoder2_relu_1” | ReLU | ReLU |
| 12 | “encoder2_conv2” | Convolution | 64 3× 3 × 64 convolutions with stride [1 1] and padding [1 1 1 1] |
| 13 | “encoder2_bn_2” | Batch normalization | Batch normalization with 64 channels |
| 14 | “encoder2_relu_2” | ReLU | ReLU |
| 15 | “encoder2_maxpool” | Max pooling | 2 × 2 max pooling with stride [2 2] and padding [0 0 0 0] |
| 16 | “decoder2_unpool” | Max unpooling | Max unpooling |
| 17 | “decoder2_conv2” | Convolution | 64 3× 3 × 64 convolutions with stride [1 1] and padding [1 1 1 1] |
| 18 | “decoder2_bn_2” | Batch normalization | Batch normalization with 64 channels |
| 19 | “decoder2_relu_2” | ReLU | ReLU |
| 20 | “decoder2_conv1” | Convolution | 64 3 × 3 × 64 convolutions with stride [1 1] and padding [1 1 1 1] |
| 21 | “decoder2_bn_1” | Batch normalization | Batch normalization with 64 channels |
| 22 | “decoder2_relu_1” | ReLU | ReLU |
| 23 | “decoder1_unpool” | Max unpooling | Max unpooling |
| 24 | “decoder1_conv2” | Convolution | 64 3 × 3 × 64 convolutions with stride [1 1] and padding [1 1 1 1] |
| 25 | “decoder1_bn_2” | Batch normalization | Batch normalization with 64 channels |
| 26 | “decoder1_relu_2” | ReLU | ReLU |
| 27 | “decoder1_conv1” | Convolution | 4 3 × 3 × 64 convolutions with stride [1 1] and padding [1 1 1 1] |
| 28 | “decoder1_bn_1” | Batch normalization | Batch normalization with 4 channels |
| 29 | “decoder1_relu_1” | ReLU | ReLU |
| 30 | “Softmax” | Softmax | Softmax |
| 31 | “Pixel_classify” | Pixel classification layer | Class weighted cross-entropy loss with “background,” “CSF,” “GM,” and “WM” classes |
Figure 3Representation of the proposed method, the training and testing image is separated, so that training images after preprocessing feeds into CNN network along with its respective ground truth, after reaching to convergence, the training is stopped and the network is now called trained. This trained network is used to test other test image separately to get segmented image, which is compared with its ground truth itself for performance analysis.
Training accuracy, intersection over union (IoU), and MeanBFscore for each assigned class.
| Accuracy | IoU | MeanBFScore | |
|---|---|---|---|
| Background | 0.98877 | 0.9855 | 0.99456 |
| CSF | 0.92848 | 0.66983 | 0.8923 |
| Gray | 0.77666 | 0.66022 | 0.93897 |
| White | 0.83603 | 0.79073 | 0.90347 |
Figure 4Confusion matrix of the experiment. The diagonal represents the accuracy of predicted class versus true class.
Figure 5Original image and ground truth image presented along with other images, as a result of proposed segmentation: (a) original image, (b) ground truth, (c) segmented image by proposed method (color), (d) CSF part (binary image), (e) GM part (binary image), (f) WM part (binary image), and (g) segmented Image in gray (c).
Comparison of performance parameters for each result image (Figure 5(g)), with respective ground truth image (Figure 5(b)).
| Test image ID | Parameter | CSF part | Gray part | White part | Mean value |
|---|---|---|---|---|---|
| OAS1_0081_MR1 | Dice similarity | 0.54 | 0.75 | 0.85 | 0.71 |
| Jaccard similarity | 0.37 | 0.59 | 0.74 | 0.57 | |
| Mean squared error | — | — | — | 29.47 | |
|
| |||||
| OAS1_0083_MR1 | Dice similarity | 0.84 | 0.75 | 0.79 | 0.80 |
| Jaccard similarity | 0.73 | 0.60 | 0.66 | 0.66 | |
| Mean squared error | — | — | — | 19.32 | |
|
| |||||
| OAS1_0084_MR1 | Dice similarity | 0.85 | 0.71 | 0.78 | 0.78 |
| Jaccard similarity | 0.74 | 0.55 | 0.64 | 0.64 | |
| Mean squared error | — | — | — | 25.02 | |
|
| |||||
| OAS1_0085_MR1 | Dice similarity | 0.72 | 0.67 | 0.73 | 0.71 |
| Jaccard similarity | 0.56 | 0.51 | 0.57 | 0.55 | |
| Mean squared error | — | — | — | 32.52 | |
|
| |||||
| OAS1_0086_MR1 | Dice similarity | 0.74 | 0.85 | 0.92 | 0.84 |
| Jaccard similarity | 0.59 | 0.74 | 0.85 | 0.73 | |
| Mean squared error | — | — | — | 9.52 | |
|
| |||||
| OAS1_0087_MR1 | Dice similarity | 0.64 | 0.75 | 0.85 | 0.74 |
| Jaccard similarity | 0.47 | 0.60 | 0.74 | 0.60 | |
| Mean squared error | — | — | — | 27.58 | |
Comparison of deep learning approaches for brain structure segmentation.
| Authors | CNN style | Dimension | Accuracy | Data |
|---|---|---|---|---|
| Zhang et al. [ | Patchwise | 2D | DSC 83.5% (CSF), 85.2% (GM), 86.4% (WM) | Private data (10 healthy infants) |
| Nie et al. [ | Semantic-pixelwise | 2D | DSC 85.5% (CSF), 87.3% (GM), 88.7% (WM) | Private data (10 healthy infants) |
| de Brebisson et al. [ | Patchwise | 2D/3D | Overall DSC 72.5% ∓ 16.3% | MICCAI 2012-multi-atlas labeling |
| Moeskops et al. [ | Patchwise | 2D/3D | Overall DSC 73.53% | MICCAI 2012-multi-atlas labeling |
| Proposed method | Pixel-label semantic (SegNet CNN) | 2D | DSC 72.2% (CSF), 74.6% (GM), 81.9% (WM) | OASIS cross-sectional MRI |