| Literature DB >> 32183041 |
Kashan Zafar1, Syed Omer Gilani1, Asim Waris1, Ali Ahmed1, Mohsin Jamil2, Muhammad Nasir Khan3, Amer Sohail Kashif1.
Abstract
Clinical treatment of skin lesion is primarily dependent on timely detection and delimitation of lesion boundaries for accurate cancerous region localization. Prevalence of skin cancer is on the higher side, especially that of melanoma, which is aggressive in nature due to its high metastasis rate. Therefore, timely diagnosis is critical for its treatment before the onset of malignancy. To address this problem, medical imaging is used for the analysis and segmentation of lesion boundaries from dermoscopic images. Various methods have been used, ranging from visual inspection to the textural analysis of the images. However, accuracy of these methods is low for proper clinical treatment because of the sensitivity involved in surgical procedures or drug application. This presents an opportunity to develop an automated model with good accuracy so that it may be used in a clinical setting. This paper proposes an automated method for segmenting lesion boundaries that combines two architectures, the U-Net and the ResNet, collectively called Res-Unet. Moreover, we also used image inpainting for hair removal, which improved the segmentation results significantly. We trained our model on the ISIC 2017 dataset and validated it on the ISIC 2017 test set as well as the PH2 dataset. Our proposed model attained a Jaccard Index of 0.772 on the ISIC 2017 test set and 0.854 on the PH2 dataset, which are comparable results to the current available state-of-the-art techniques.Entities:
Keywords: Jaccard Index; ROC curve; ResNet; U-Net; convolutional neural networks; dermoscopic images; image inpainting; melanoma
Mesh:
Year: 2020 PMID: 32183041 PMCID: PMC7147706 DOI: 10.3390/s20061601
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Examples of the ISIC-17 dataset (a) and PH2 dataset (b).
Figure 2Representation of a single image as it is passed through the hair removal algorithm. Left to right: (a) Input image; (b) Grayscale image; (c) Cross shaped structuring element employed during morphological operations; (d) Image obtained after applying black top-hat filter; (e) Image after thresholding; (f). Final image obtained as output.
Figure 3Schematic diagram representing UResNet-50. The encoder shown on the left is the ResNet-50, while the U-Net decoder is shown on the right. Given in the parenthesis is the channel dimensions of the incoming feature maps to each block. Arrows are defined in the legend.
Convolutional network architecture based on ResNet-50.
| Layer Name | Output Size | Kernel Size & No. of Filters |
|---|---|---|
|
| 128 × 128 | 7 × 7, 64, Stride 2 |
|
| 64 × 64 | 3 × 3, Stride 2 |
|
| 64 × 64 | |
|
| 32 × 32 | |
|
| 16 × 16 | |
|
| 8 × 8 |
The deconvolution architecture based on U-Net.
| Layer Name | Kernel | Output Size & No. of Filters |
|---|---|---|
| U1 | 2 × 2 | 16 × 16 × 2048 |
| D1 | 3 × 3 | 16 × 16 × 256 |
| D2 | 3 × 3 | 16 × 16 × 256 |
| U2 | 2 × 2 | 32 × 32 × 256 |
| D3 | 3 × 3 | 32 × 32 × 128 |
| D4 | 3 × 3 | 32 × 32 × 128 |
| U3 | 2 × 2 | 64 × 64 × 128 |
| D5 | 3 × 3 | 64 × 64 × 64 |
| D6 | 3 × 3 | 64 × 64 × 64 |
| U4 | 2 × 2 | 128 × 128 × 64 |
| D7 | 3 × 3 | 128 × 128 × 32 |
| D8 | 3 × 3 | 128 × 128 × 32 |
| U5 | 2 × 2 | 256 × 256 × 32 |
| D9 | 3 × 3 | 256 × 256 × 16 |
| D10 | 3 × 3 | 256 × 256 × 16 |
| Output | 1 × 1 | 256 × 256 × 1 |
Hyperparameters maintained during training.
| Name | Value |
|---|---|
| Input Size | 256 × 256 × 3 |
| Batch Size | 16 |
| Learning Rate | 1 × 10−3 |
| Optimizer | Adam |
| Epoch | 100 |
| Loss Function | Binary Crossentropy |
Figure 4Training and validation accuracy of the proposed convolutional neural network model for 70 epochs.
Figure 5Example results of multiple patients. The first row contains the original images of five patients from the test set. The second row contains corresponding ground truths as provided. The third row contains predicted masks from the proposed method.
Figure 6Receiver operative characteristics (ROC) curve generated on the ISIC-17 test set.
Model performance on the ISIC-2017 test set.
| Methods | Jaccard Index |
|---|---|
| No Preprocessing | Preprocessed |
| 0.763 | 0.772 |
Comparison with different frameworks on the ISIC-2017 test set.
| Methods | Jaccard Index | Dice Coefficient |
|---|---|---|
| FCN-8s [ | 0.696 | 0.783 |
| U-Net [ | 0.651 | 0.768 |
| II-FCN [ | 0.699 | 0.794 |
| Auto-ED [ | 0.738 | 0.824 |
| FCRN [ | 0.753 | 0.839 |
| Res-Unet (Proposed) | 0.772 | 0.858 |
Comparison of results with the challenge participants.
| Authors | Model | Jaccard Index |
|---|---|---|
| Yading Yuan et al. [ | CDNN | 0.765 |
| Matt Berseth et al. [ | U-Net | 0.762 |
| Popleyi et al. [ | FCN | 0.760 |
| Euijoon Ahn et al. [ | ResNet | 0.758 |
| Afonso Menegola et al. [ | VGG16 | 0.754 |
| Proposed Method | Res-Unet | 0.772 |
Comparison with different frameworks on the PH2 Dataset
| Methods | Jaccard Index | Dice Coefficient |
|---|---|---|
| FCN-16s | 0.802 | 0.881 |
| DeeplabV3+ | 0.814 | 0.890 |
| Mask-RCNN | 0.830 | 0.904 |
| Multi-Stage FCN | - | 0.906 |
| SSLS | - | 0.916 |
| Res-Unet (Proposed) | 0.854 | 0.924 |