| Literature DB >> 35273282 |
Hassan Ashraf1, Asim Waris2, Muhammad Fazeel Ghafoor1, Syed Omer Gilani1, Imran Khan Niazi3,4,5.
Abstract
In a computer-aided diagnostic (CAD) system for skin lesion segmentation, variations in shape and size of the skin lesion makes the segmentation task more challenging. Lesion segmentation is an initial step in CAD schemes as it leads to low error rates in quantification of the structure, boundary, and scale of the skin lesion. Subjective clinical assessment of the skin lesion segmentation results provided by current state-of-the-art deep learning segmentation techniques does not offer the required results as per the inter-observer agreement of expert dermatologists. This study proposes a novel deep learning-based, fully automated approach to skin lesion segmentation, including sophisticated pre and postprocessing approaches. We use three deep learning models, including UNet, deep residual U-Net (ResUNet), and improved ResUNet (ResUNet++). The preprocessing phase combines morphological filters with an inpainting algorithm to eliminate unnecessary hair structures from the dermoscopic images. Finally, we used test time augmentation (TTA) and conditional random field (CRF) in the postprocessing stage to improve segmentation accuracy. The proposed method was trained and evaluated on ISIC-2016 and ISIC-2017 skin lesion datasets. It achieved an average Jaccard Index of 85.96% and 80.05% for ISIC-2016 and ISIC-2017 datasets, when trained individually. When trained on combined dataset (ISIC-2016 and ISIC-2017), the proposed method achieved an average Jaccard Index of 80.73% and 90.02% on ISIC-2017 and ISIC-2016 testing datasets. The proposed methodological framework can be used to design a fully automated computer-aided skin lesion diagnostic system due to its high scalability and robustness.Entities:
Mesh:
Year: 2022 PMID: 35273282 PMCID: PMC8913825 DOI: 10.1038/s41598-022-07885-y
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1The architectures of utilized UNet, ResUNet and ResUNet++ models. All three architectures have encoder, bridge, and decoder blocks. In all these blocks 3 3 convolutional layers has been used except in output block, where a 1 1 convolutional layer followed by a sigmoid function is used to convert segmentation map into the predicted mask. Combinedly, 21, 15 and 41 convolutional layers has been used in UNet, ResUNet and ResUNet++ architectures, respectively.
Figure 2The building blocks of neural network. (a) shows a plain neural unit used in UNet, whereas (b) shows a residual unit used in ResUNet and ResUNet++ architectures.
Figure 3Some example dermoscopic images from ISIC-2016 and ISIC-2017 datasets. The images of both datasets vary in characteristic making the segmentation process more challenging.
Figure 4The qualitative results of the proposed hair removing technique in the preprocessing stage. The raw RGB dermoscopic image is first converted to grayscale, then a black top-hat filter followed by a thresholding operation is applied. Finally, by utilizing inpainting algorithm the hair structures are removed while preserving the shape of the image.
Different type of pixel and spatial level image transformations used in this study.
| Augmentation type | Augmentation |
|---|---|
| Pixel level | Motion blur |
| Random brightness | |
| Random contrast | |
| Random gamma | |
| Hue saturation value | |
| RGB shift | |
| Grayscale | |
| Spatial level | Random 90 rotation |
| Vertical flip | |
| Horizontal flip |
Different type of loss functions with description used in this study.
| Loss function | Description |
|---|---|
| Binary cross-entropy | |
| Dice loss | |
| Binary cross entropy—Dice loss | |
| Focal loss | BCE is used to derive Focal Loss (F.L.), cross-entropy function can be rewritten as: Furthermore, the F.L. is calculated using a modulating factor |
| Focal Tversky loss | |
| Tversky loss | Here, |
Performance evaluation metrics used to evaluate the performance of proposed skin lesion segmentation framework.
| Performance metric | Description |
|---|---|
| Precision | |
| Recall | |
| Jaccard index | |
| Dice coefficient |
Effect of different augmentation schemes employed in this study. The models were first trained and tested without using train-time augmentations, in next experiments different combinations were used to assess the effect of image augmentations on the performance of proposed technique.
| Augmentations | Model | Jaccard index |
|---|---|---|
| No augmentations | UNet | 82.34 |
| ResUNet | 79.22 | |
| ResUNet++ | 82.87 | |
| Rotation, vertical, and horizontal flip | UNet | 82.25 |
| ResUNet | 79.28 | |
| ResUNet++ | 83.01 | |
| Rotation, vertical,horizontal flip, hue saturation value, RGB shift random brightness, and random contrast | UNet | 83.39 |
| ResUNet | 80.35 | |
| ResUNet++ | 83.44 | |
| Rotation, vertical,horizontal flip, grayscale hue saturation value, RGB shift, random brightness random contrast, motion blur, and random gamma | UNet | 82.72 |
| ResUNet | 79.81 | |
| ResUNet++ | 83.10 |
The effect of different input image sizes on the performance of proposed methodology.
| Method | Model | Jaccard index |
|---|---|---|
| 128 | UNet | 80.35 |
| ResUNet | 77.10 | |
| ResUNet++ | 81.33 | |
| 192 | UNet | 82.60 |
| ResUNet | 79.56 | |
| ResUNet++ | 82.49 | |
| 256 | UNet | 83.39 |
| ResUNet | 80.35 | |
| ResUNet++ | 83.21 | |
| 512 | UNet | 83.12 |
| ResUNet | 80.40 | |
| ResUNet++ | 83.44 |
Figure 5Effect of different loss functions on the performance of proposed semantic segmentation techniques. In all conducted experiments, for all three neural networks, Dice Loss achieved the highest Jaccard Index.
Effect of batch size on the performance of proposed skin lesion segmentation system.
| Method | Model | Jaccard index |
|---|---|---|
| 16 | UNet | 84.11 |
| ResUNet | 80.75 | |
| ResUNet++ | 83.44 | |
| 32 | UNet | 83.39 |
| ResUNet | 80.35 | |
| ResUNet++ | 82.67 | |
| 64 | UNet | 82.32 |
| ResUNet | 79.03 | |
| ResUNet++ | 82.11 |
Effect of preprocessing on the performance of proposed skin lesion segmentation system.
| Method | Model | Jaccard index |
|---|---|---|
| With preprocessing | UNet | 82.84 |
| ResUNet | 79.34 | |
| ResUNet++ | 82.45 | |
| Without preprocessing | UNet | 81.39 |
| ResUNet | 79.15 | |
| ResUNet++ | 81.88 |
Effect of postprocessing technique on the performance of proposed skin lesion segmentation system.
| Model | Normal | CRF | TTA | CRF-TTA |
|---|---|---|---|---|
| UNet | 83.84 | 83.94 | 84.59 | 84.60 |
| ResUNet | 80.34 | 80.59 | 80.89 | 81.07 |
| ResUNet++ | 82.45 | 82.44 | 82.43 | 82.44 |
Figure 6Qualitative performance of proposed skin lesion segmentation technique on ISIC-2016 and ISIC-2017 datasets. The first column shows the original testing images, the actual annotated masks are shown in column 2. The third column shows the predicted mask without using any postprocessing technique. In fourth and fifth column the predicted masks refined by CRF and TTA are shown. The last column shows the predicted mask refined by using both CRF and TTA.
Results of proposed methodology trained and tested on combined dataset.The results presented in this table are resulted from the models trained on combined dataset.
| Method | Dice | Jaccard | Precision | Recall |
|---|---|---|---|---|
| UNet | 87.26 | 84.60 | 89.18 | 86.10 |
| ResUNet | 83.59 | 81.07 | 84.61 | 85.06 |
| ResUNet++ | 83.73 | 85.44 | 86.11 | 93.85 |
| UNet | 93.74 | 89.59 | 95.03 | 90.21 |
| ResUNet | 90.70 | 86.64 | 91.38 | 89.31 |
| ResUNet++ | 92.71 | 90.02 | 94.34 | 92.19 |
| UNet | 83.74 | 80.65 | 85.00 | 85.15 |
| ResUNet | 79.83 | 77.89 | 81.33 | 83.37 |
| ResUNet++ | 82.43 | 80.73 | 86.37 | 87.01 |
Figure 7Pixel level ROCs resulted from proposed technique corresponding to all three utilized neural networks. The first row shows the area under the curves (AUCs) for the model trained and tested on combined dataset. the second and third row illustrated the AUCs for the models trained on combined data and tested on ISIC-2016 and ISIC-2017 dataset, respectively.
Comparison of proposed methodology with state-of-the-art techniques and competition participants.The results presented in this table are resulted from models trained and tested on individual ISIC-2016 and ISIC-2017 datasets.
| Method | Dice | Jaccard | Precision | Recall |
|---|---|---|---|---|
| Team-EXB[ | 91.00 | 84.30 | 96.50 | 91.00 |
| Team-CUMED[ | 89.70 | 82.90 | 91.10 | 95.70 |
| Team-Rahman[ | 89.50 | 82.22 | 88.00 | 96.90 |
| Bi et al.[ | 91.18 | 84.64 | 92.17 | 96.54 |
| Yuan et al.[ | 91.20 | 84.70 | 96.60 | 91.80 |
| Lesion Net[ | 92.39 | 96.45 | 93.62 | |
| 90.39 | 85.96 | 87.46 | 92.76 | |
| 88.33 | 84.15 | 84.75 | 92.22 | |
| 90.27 | 85.88 | 93.60 | 86.27 | |
| Yuan et el. (CDNN)[ | 84.90 | 76.50 | 97.50 | 82.50 |
| Li et al.[ | 84.70 | 76.20 | 97.80 | 82.00 |
| Bi et al. (ResNet)[ | 84.40 | 76.00 | 98.50 | 80.20 |
| Lin et al. (UNet)[ | 77 | 62.00 | – | – |
| Al-masni et al. (FrCN)[ | 87.08 | 77.11 | 96.69 | 85.40 |
| Kashan et al. (ResUNet)[ | 85.80 | 77.20 | – | – |
| Lesion Net[ | 87.87 | 78.28 | 96.08 | 86.23 |
| 90.39 | 86.77 | 82.17 | ||
| 79.28 | 77.66 | 82.46 | 80.47 | |
| 82.96 | 80.03 | 84.01 | 85.26 | |
Significant values are in bold.
Figure 8Distribution of jaccard index on ISIC-2016 and ISIC-2017 dataset. Each bar is labelled with number of images falling in the bin range of bars.
Figure 9Some exemplary segmented images from ISIC-2017 dataset. The proposed skin lesion technique can handle the dermoscopic images acquired in different conditions. The model performed well on images with low contrast, invisible lesions, and images with hair structures.
Figure 10Failure cases in ISIC-2017 dataset. The provided annotated ground truth masks of some images are annotated incorrectly (subjective assessment). The model still predicted the masks quite resembled with the actual lesions, but the IoU score was low since it was assigned on the basis of provided versus predicted mask.