Literature DB >> 35161613

Modified U-NET Architecture for Segmentation of Skin Lesion.

Vatsala Anand1, Sheifali Gupta1, Deepika Koundal2, Soumya Ranjan Nayak3, Paolo Barsocchi4, Akash Kumar Bhoi4,5,6.   

Abstract

Dermoscopy images can be classified more accurately if skin lesions or nodules are segmented. Because of their fuzzy borders, irregular boundaries, inter- and intra-class variances, and so on, nodule segmentation is a difficult task. For the segmentation of skin lesions from dermoscopic pictures, several algorithms have been developed. However, their accuracy lags well behind the industry standard. In this paper, a modified U-Net architecture is proposed by modifying the feature map's dimension for an accurate and automatic segmentation of dermoscopic images. Apart from this, more kernels to the feature map allowed for a more precise extraction of the nodule. We evaluated the effectiveness of the proposed model by considering several hyper parameters such as epochs, batch size, and the types of optimizers, testing it with augmentation techniques implemented to enhance the amount of photos available in the PH2 dataset. The best performance achieved by the proposed model is with an Adam optimizer using a batch size of 8 and 75 epochs.

Entities:  

Keywords:  U-Net; convolutional neural network; deep learning; image analysis; segmentation; skin disease

Mesh:

Year:  2022        PMID: 35161613      PMCID: PMC8838042          DOI: 10.3390/s22030867

Source DB:  PubMed          Journal:  Sensors (Basel)        ISSN: 1424-8220            Impact factor:   3.576


1. Introduction

All kinds of microorganisms can cause a variety of skin infections, which can range from moderate to serious. An early diagnosis of a skin illness improves the chances of a successful treatment. Dark patches or birthmarks may appear on the skin as a result of skin illness [1]. The modality used for the diagnosis of skin disease is dermoscopy images. Dermoscopy is a process in which a dermatologist observes a position with a special microscope or magnifying lens. Dermatologist uses a device named dermatoscope, that consists of a high magnifying class, so that a clear picture of the nodule can be seen. In medical imaging, a variety of technologies are employed for the diagnosis of skin disease such as Machine Learning (ML), Deep Learning (DL), Transfer Learning (TL), Ensemble Learning (EL), etc. In ML, a machine is made to learn the tasks, whereas in DL, TL, and EL it learns features directly from the data provided. Improvements in deep learning Convolutional Neural Networks (CNN) have shown promising results in recent years, and they have also emerged as a difficult study topic for categorization in medical image processing [2,3]. Automatically segmenting skin lesions is still a difficult task. Some skin lesions with light pigment have a very similar color and visual patterns in the pigment patches and the surrounding skin regions, making skin lesion segmentation difficult. In addition, the original dermoscopic images have a high resolution, which means that processing them on computers takes a lot of time and resources [4]. All of these factors contribute to the color and texture distribution of the lesions and impede successful learning in the skin lesions themselves. Certain pictures also show hairs and col-or-makers, making skin lesion segmentation more difficult. The segmentation of skin lesions is challenging due to these difficulties. However, recent studies have shown interest on different variations of DL and CNN models to overcome the above-mentioned difficulties that arise in segmentation [5,6,7,8,9]. Indeed, Yuan et al. [10] presented a completely automated technique for skin lesion segmentation based on 19-layer deep convolutional neural networks that are trained end-to-end and do not require previous data knowledge. They have obtained a value of Jaccard distance of 0.963. In 2017, Yuan et al. [11] proposed convolutional and deconvolutional neural networks using ISBI challenge and obtained a value of Jaccard distance of 0.784. Schaefer et al. [12] combined the enhancement stage with two segmentation algorithms. In the first algorithm, they derived the optimal value of threshold, and in the second algorithm they applied lesion edge detection. Various segmentation techniques developed in skin lesions are shown in literature [13,14]. Shankar et al. [15] showed a histogram thresholding approach to determine the segmented threshold values from the tissues. Pustokhina et al. [16] showed the advantages of edge and region-based techniques that include edge operations and models such as region splitting and merging, etc. Raj et al. [17] used supervised models such as Random forest for skin lesion segmentation by training the model. Anand et al. [18] used pre-processing, segmentation, and classification for the diagnosis of disease. Moreover, they showed that in medical images, the presence of fuzziness and ambiguity produces a wide range of outcomes. Garnavi et al. [19] presented a hybrid threshold-based border detection method to determine nodule areas. Ganster et al. [20] proposed an integrated segmentation technique that is based on thresholding and clustering algorithms. Erkol et al. [21] presented a segmentation method based on the Laplacian-of-Gaussian and active contour methods. Moreover, in some of the literature, authors have classified skin lesions directly without performing any segmentation of the ROI part. For skin disease classification, various features were used according to the ABCDE rule [22,23]. According to the ABCDE rule, different features such as Asymmetry, Border detection, Color, Diameter and Evolving are used. Recently, UNet architecture was proposed for segmenting lesions and their attributes based on the CNN architecture. A 100-image validation set and a 1000-image test set were derived from ISIC 2018 [24]. Masni et al. [25] used a total of 2750 images from the ISIC’s 2016, 2017, and 2018 datasets. They worked only on three classes—NV, MEL, and AKIEC—and on four pre-trained models—Inception-v3, ResNet-50, Inception-ResNet-v2, and-DenseNet201—and achieved an accuracy of 81.29%, 81.57%, 81.34%, and 73.44%, respectively. Dorj et al. [26] concentrated on utilizing CNN to classify skin cancer. For feature extraction, they employed a pre-trained AlexNet model and used a total of 3753 images. The authors used various classification models for the diagnosis in health applications [27,28,29], and the machine learning methods with ensemble are used with EEG signal [30,31,32,33]. Segmented images of skin lesions have been used by a few researchers to increase classification accuracy. However, image classification, segmentation, and object recognition have all benefited from the increased attention given to Deep Learning (DL) models [34]. In this context, Hang Li et al. [35] proposed dense deconvolutional network on ISBI 2016 and 2017 datasets for skin lesion segmentation based on residual learning. The proposed dense deconvolutional networks reused the learned features from the previous layers, which establishes dense links among all feature maps. By addressing the issue of gradient vanishing, the proposed method strengthens the propagation of multi-level features in the entire network and boosts the performance of skin lesion segmentation greatly. On the 2016 dataset, they obtained a 0.870 value of Jaccard coefficient, and on the 2017 dataset they obtained a 0.765 value of Jaccard coefficient. Yu et al. [36] developed a segmentation and deep classification network with two tiers. They developed a fusion network and obtained an accuracy value of 86.54. Khan et al. [37] proposed a model using pre-processing, segmentation and classification and obtained a segmentation accuracy of 96.8% and 92.1% for the ISIC and PH2 datasets, respectively. Moreover, they obtained a classification accuracy of 97% on the ISIC dataset. Long et al. [38] used the concept of fine tuning with the classification networks AlexNet, GoogleNet, and VGGNet. The authors presented a novel architecture that produces accurate and detailed results by combining semantic information from a deep, coarse layer with appearance information from a shallow, fine layer’s segmentations. Chen et al. [39] combined the methods from deep convolutional neural networks and graphical models for addressing semantic image segmentation. The authors combined the final layers with a fully connected Conditional Random Field. They obtained an IOU accuracy value of 71.6% in the test set. Noh et al. [40] combined a deep deconvolution network, and the proposed technique overcomes the constraints of previous methods based on fully convolutional networks, and the segmentation method frequently finds complex structures and handles objects at many sizes. They achieved an accuracy of 72.5% through ensembling with the fully convolutional network. Wang et al. [41] presented non-local U-Nets that are used with flexible global aggregation blocks. These blocks are used for preserving size in upsampling and downsampling layers. Ibethaz et al. [42] proposed a MultiResUNet architecture, by replacing it with the two convolutional layers. A parameter is assigned for the layers that controls the number of filters of the convolutional layers. They used 97 images ranging from 1349 × 1030 to 1344 × 1024 and resized to 256 × 256, and used the ISIC-2018 dataset. Christ et al. [43] presented a method for the automatic segmentation of lesions in CT abdomen images using cascaded fully convolutional neural networks. They used a two-fold cross validation on the images and obtained a dice score of over 94%. Lin et al. [44] presented a generic multi-path refinement network that explicitly exploits all the information available along the down-sampling process to enable a high-resolution prediction using long-range residual connections. They achieved an intersection-over-union score of 83.4 on the challenging PASCAL VOC 2012 dataset. Novikov et al. [45] used a multi-class configuration, and the ground-truth masks were trained and tested using the publicly accessible JSRT database, which contains 247 X-ray images and can be found in the SCR database. They obtained a Jaccard index value of 95.0%. The performance results after segmentation increase, and the results obtained are satisfying. From the literature, it can be seen that, when the segmented images are used for classification, the classification accuracy increases. The major contributions of the study are as follows: A modified U-Net architecture has been proposed for the segmentation of lesions from skin disease using dermoscopy images. The data augmentation technique has been performed to increase the randomness of images for better stability. The proposed model is validated with different optimizers, batch sizes, and epochs for better accuracy. The proposed model has been analyzed with various performance parameters such as Jaccard Index, Dice Coefficient, Precision, Recall, Accuracy and Loss. The rest of the paper is structured as follows: materials and methods are given in Section 2, followed by results and discussions in Section 3, and Section 4 shows the conclusion and future scope.

2. Materials and Methods

The proposed model exploits the U-Net architecture for lesion segmentation from skin disease dermoscopy images. The proposed model has been evaluated on the PH2 [46] dataset consisting of 200 skin disease dermoscopy images.

2.1. Dataset

The PH2 dataset contains 200 dermoscopy images (160 non-melanomas and 40 melanoma) that are obtained by the Tuebinger Mole Analyzer system using a 20-fold magnification. All images have an approximate size of 192 × 256 pixels. Figure 1 shows the skin diseases’ original images, and Figure 2 shows the ground truth masks for the respective original images.
Figure 1

Skin Disease Original Images: (a) Image 1; (b) Image 2; (c) Image 3; (d) Image 4.

Figure 2

Ground Truth Masks for Respective Original Images: (a) Image 1; (b) Image 2; (c) Image 3; (d) Image 4.

2.2. Data Augmentation

As the available training dermoscopy images in the dataset are few, offline data augmentation techniques have been implemented to increase the number of sample images. The data augmentation [47] on images is done using different techniques such as flipping and rotation, as shown in Figure 3. The corresponding masks of the augmented images are also shown in Figure 3.
Figure 3

Images and Respective Masks after Data Augmentation Techniques: (a) Original Image; (b) Rotated Image; (c) Flipped Image; (d) Original Mask; (e) Rotated Mask; (f) Flipped Mask.

2.3. Modified U-Net Architecture

An enhanced version of the Convolutional Neural Network (CNN) was developed for dealing with biomedical images in which the purpose is not only to categorize whether or not an infection exists but also to identify the infected area [48]. The U-Net architecture consists of two paths. The first one is the contraction path, that is also known as encoder, and the second one is the symmetric expanding path, also known as decoder. Encoder is used to capture the image context, whereas decoder uses transposed convolutions to enable precise localization. In this paper, the proposed modified U-Net architecture has been presented, as shown in Figure 4.
Figure 4

Modified U-Net Architecture.

The proposed architecture localizes and distinguishes borders by classifying every pixel; therefore, input and output share the same size. In the encoder part, the convolution layer and the max-pooling layer are applied. In the decoder part, the transposed convolution layer and the simple convolution layer are applied. During the simulation phase, the Input image undergoes a multilevel decomposition in the encoder path, and the feature maps are reduced with the help of a max pooling layer, which can be seen in Figure 4 as arrows with different colors. The yellow arrows show the convolutional layer of size 3 × 3, ReLU (Rectified Linear Unit) activation function and dropout layer; the red arrows show the convolutional layer of size 3 × 3 and ReLU activation function; the blue arrows show the max-pooling layer; the green arrows show the upsampling with 2 × 2 size; the black arrows show the concatenation of images from contracting and expanding paths; and, finally, the brown arrows show the final convolutional layer with size 1 × 1. In the contraction path, each and every process consists of two convolutional layers. In the first part, the channel changes from 1 to 64. The blue arrow pointing down shows the max pooling layer that halves down the image from 192 × 256 to 96 × 128. This process is repeated three times and reaches below. Below are the two convolutional layers, but these layers are without max pooling layers. The image has been resized to 12 × 16 × 1024. In the expanding path, the image is going to be upsized to its original image size. The upsampling technique expands the size of the images, and it is known as transposed convolution. The image is upsized from 12 × 16 to 24 × 32. After that, the image is concatenated with the image from the contracting path. The reason for this is to combine the information from the last layers to get a more accurate prediction. The proposed modified U-Net architecture includes a feature map rectangular in size starting from 192 × 256 in the first layer and 96 × 128 in the second layer. It is downsized again to 48 × 64 in the third layer. Then, it is downsized to 24 × 32 in the fourth layer, and, finally, it is downsized to 6 × 8 in the last layer. Afterwards, the feature map size increases in the expansion path with 24 × 32 in the first layer from the bottom. It is upsized to 48 × 64 in the second layer and to 96 × 128 in the third layer. Finally, the feature map size changes to 192 × 256 in the topmost layer. After the contraction and expanding process, the architecture reaches the upper level, reshaping the image; the last layer is a convolution layer. Table 1 shows the parameters of the proposed model, that consists of different convolution layers, input and output image size, filter size, number of filters, and activation function. The total number of parameters for the proposed model are 33,393,669, whereas the total number of trainable parameters are 33,377,795, and non-trainable parameters are 15,874.
Table 1

Parameters of the Proposed Model.

S. No.LayersInput Image SizeFilter SizeNo. of FilterActivation FunctionOutput Image SizeParameters
1Input Image192 × 256 × 3-----
2Conv_1192 × 256 × 33 × 364ReLU192 × 256 × 641792
3Batch Normalization192 × 256 × 64----256
4Conv 2192 × 256 × 33 × 364ReLU192 × 256 × 6436,928
5Batch Normalization192 × 256 × 64----256
6MaxPooling192 × 256 × 643 × 3--96 × 128 × 640
7Conv_396 × 128 × 1283 × 3128ReLU96 × 128 × 12873,856
8Batch Normalization96 × 128 × 128----512
9Conv 496 × 128 × 1283 × 3128ReLU96 × 128 × 128147,584
10Batch Normalization96 × 128 × 128----512
11MaxPooling96 × 128 × 1283 × 3--48 × 64 × 1280
12Conv 548 × 64 × 2563 × 3256ReLU48 × 64 × 256295,168
13Batch Normalization48 × 64 × 256----1024
14Conv 648 × 64 × 2563 × 3256ReLU96 × 128 × 128590,080
15Batch Normalization48 × 64 × 256----1024
16MaxPooling48 × 64 × 2563 × 3--48 × 64 × 128
17Conv 748 × 64 × 2563 × 3256ReLU96 × 128 × 128590,080
18Batch Normalization48 × 64 × 256----1024
19MaxPooling48 × 64 × 2563 × 3--24 × 32 × 2560
20Conv 824 × 32 × 5123 × 3512ReLU24 × 32 × 5121,180,160
21Batch Normalization24 × 32 × 512----2048
22Conv 924 × 32 × 5123 × 3512ReLU24 × 32 × 5122,359,808
23Batch Normalization24 × 32 × 512----2048
24Conv 1024 × 32 × 5123 × 3512ReLU24 × 32 × 5122,359,808
25Batch Normalization24 × 32 × 512----2048
26MaxPooling24 × 32 × 5123 × 3--12 × 16 × 5120
27Conv 1112 × 16 × 5123 × 3512ReLU12 × 16 × 5122,359,808
28Batch Normalization12 × 16 × 512----2048
29Conv 1212 × 16 × 5123 × 3512ReLU12 × 16 × 5122,359,808
30Batch Normalization12 × 16 × 512----2048
31Conv 1312 × 16 × 5123 × 3512ReLU12 × 16 × 5122,359,808
32Batch Normalization12 × 16 × 512----2048
33MaxPooling12 × 16 × 5123 × 3--6 × 8 × 5120
34Upsampling12 × 16 × 1024---12 × 16 × 10240
35De-Conv 112 × 16 × 5123 × 3512ReLU12 × 16 × 5124,719,104
36Batch Normalization12 × 16 × 512----2048
37De-Conv 212 × 16 × 5123 × 3512ReLU12 × 16 × 5122,359,808
38Batch Normalization12 × 16 × 512----2048
39De-Conv 312 × 16 × 5123 × 3512ReLU12 × 16 × 5122,359,808
40Batch Normalization12 × 16 × 512----2048
41Upsampling24 × 32 × 512---24 × 32 × 5120
42De-Conv 424 × 32 × 5123 × 3512ReLU24 × 32 × 5122,359,808
43Batch Normalization24 × 32 × 512----2048
44De-Conv 524 × 32 × 5123 × 3512ReLU24 × 32 × 5122,359,808
45Batch Normalization24 × 32 × 512----2048
46De-Conv 624 × 32 × 2563 × 3512ReLU24 × 32 × 5121,179,904
47Batch Normalization24 × 32 × 256----1024
48Upsampling48 × 64 × 256---48 × 64 × 2560
49De-Conv 748 × 64 × 2563 × 3512ReLU48 × 64 × 256590,080
50Batch Normalization48 × 64 × 256----1024
51De-Conv 848 × 64 × 2563 × 3512ReLU48 × 64 × 256590,080
52Batch Normalization48 × 64 × 256----1024
53De-Conv 948 × 64 × 1283 × 3512ReLU48 × 64 × 256295,040
54Batch Normalization48 × 64 × 128----512
55Upsampling96 × 128 × 128---96 × 128 × 1280
56De-Conv 1096 × 128 × 1283 × 3512ReLU96 × 128 × 128147,584
57Batch Normalization96 × 128 × 128----512
58De-Conv 1196 × 128 × 643 × 3512ReLU96 × 128 × 6473,792
59Batch Normalization96 × 128 × 64----256
60Upsampling192 × 256 × 64---192 × 256 × 640
61De-Conv 12192 × 256 × 643 × 3512ReLU192 × 256 × 6436,928
62Batch Normalization192 × 256 × 64----256
63De-Conv 13192 × 256 × 13 × 3512ReLU192 × 256 × 1577
64Batch Normalization192 × 256 × 1----4
Total Parameters = 33,393,669
Trainable Parameters = 33,377,795
Non-Trainable Parameters = 15,874

3. Results and Discussion

This section includes all the results attained by using a modified U-Net model. The model is evaluated on the PH2 dataset. An experimental analysis has been done, from which training accuracy and loss curves are obtained. A detailed description of the performed visual analysis of segmented images and the analysis of confusion matrix parameters is given below.

3.1. Result Analysis Based on Different Optimizers

This section includes all the results obtained by using Adam, Adadelta, and SGD optimizers with a batch size of 18 and 100 epochs.

3.1.1. Analysis of Training Loss and Accuracy

The results are taken using different optimizers with a batch size of 18 and 100 epochs. Figure 5 shows the curves of training loss and training accuracy. It is worth noticing that the value of accuracy increases with the number of epochs, and the loss value decreases. The color red shows the training loss, and the color blue shows the training accuracy.
Figure 5

Analysis of training loss and training accuracy: (a) Training Loss with SGD optimizer, (b) Training Accuracy with SGD optimizer, (c) Training Loss with Adam optimizer, (d) Training Accuracy with Adam optimizer, (e) Training Loss with Adadelta optimizer, (f) Training Accuracy with Adadelta optimizer’s Original Image.

Figure 5a shows the training loss by using the SGD optimizer; the maximum loss value is 0.7, which decreases with the number of epochs. Figure 5b shows the training accuracy in which the maximum accuracy is greater than 0.95 at the 100th epoch. Figure 5c shows the training loss by exploiting the Adam optimizer; the maximum loss value is lower than that of the SGD optimizer. Figure 5d shows the training accuracy in which the maximum accuracy is greater than 0.975 at the 100th epoch. The accuracy for the Adam optimizer outperforms the accuracy at the SGD optimizer. Figure 5e shows the training loss with the Adadelta optimizer, whose maximum value is 0.75, which is higher with respect to the SGD and Adam optimizers. Figure 5f shows the training accuracy, and the value of accuracy is only 0.90. Figure 5 shows that the Adam optimizer outperforms the SGD and Adadelta optimizers in terms of training loss and training accuracy.

3.1.2. Visual Analysis of Segmented Images

Figure 6 shows the segmented images using the Adam, Adadelta and SGD optimizers with a batch size of 18 and 100 epochs. Figure 6a,c shows the ground truth masks of the original images, and Figure 6b,d shows the original images. Figure 6e,g shows the predicted masks of original images 1 and 2 with the Adam optimizer, whereas Figure 6f,h shows the segmented outputs of original images 1 and 2 with the Adam optimizer. Similarly, Figure 6i–p shows the predicted masks and segmented outputs for the Adadelta and SGD optimizers, respectively. From the visual analysis of these figures, it can be seen that the Adam and SGD optimizers show almost similar results with a batch size of 18 and 100 epochs, whereas the Adadelta optimizer does not follow the profile of the skin lesion; rather, it extracts a complete skin part. So, the Adadelta optimizer cannot be recommended for skin lesion segmentation. To select the best performing optimizer between Adam and SGD, an analysis of these two optimizers is done in Section 3.1.3 using confusion matrix parameters.
Figure 6

Images segmented with a Batch Size of 18, 100 epochs and different optimizers: (a) Ground truth Mask of Original Image 1; (b) Original Image 1; (c) Ground truth Mask of Original Image 2; (d) Original Image 2; (e) Predicted Mask of Image 1 with Adam optimizer; (f) Segmented Output of Image 1 with Adam optimizer; (g) Predicted Mask of Image 2 with Adam optimizer; (h) Segmented Output of Image 2 with Adam optimizer and; (i) Predicted Mask of Image 1 with Adadelta optimizer; (j) Segmented Output of Image 1 with Adadelta optimizer; (k) Predicted Mask of Image 2 with Adadelta optimizer; (l) Segmented Output of Image 2 with Adadelta optimizer; (m) Predicted Mask of Image 1 with SGD optimizer; (n) Segmented Output of Image 1 with SGD optimizer; (o) Predicted Mask of Image 2 with SGD optimizer; (p) Segmented Output of Image 2 with SGD optimizer.

3.1.3. Analysis of Confusion Matrix Parameters

In Section 3.1.2, a visual analysis of segmented images is shown, proving that the Adam and SGD optimizers do not have the best results. Now, to see the best performing optimizer, confusion matrix parameters are analyzed. Table 2 shows the comparison of the Jaccard Index, Dice Coefficient, Precision, Recall, Accuracy, and Loss for the modified U-Net model architecture by using the Adam, Adadelta, and SGD optimizers.
Table 2

Analysis of Different Optimizers with a batch size of 18 and 100 epochs.

Training Dataset
OptimizerJaccard Index (%)Dice Coefficient (%)Precision (%)Recall (%)Accuracy (%)Loss
SGD96.8184.6096.0996.8697.7712.03
Adam96.4288.3292.1598.5096.8811.31
Adadelta83.9061.6286.4395.8293.9138.33
Testing Dataset
Jaccard Index (%) Dice Coefficient (%) Precision (%) Recall (%) Accuracy (%) Loss
SGD93.9880.2690.6091.6494.5517.91
Adam93.8384.8685.8996.9394.0419.19
Adadelta82.4159.1281.0890.8290.5541.54
Validation Dataset
Jaccard Index (%) Dice Coefficient (%) Precision (%) Recall (%) Accuracy (%) Loss
SGD94.4481.0191.2392.6594.7917.37
Adam94.7486.1388.3097.1495.0116.24
Adadelta82.6060.1380.7692.6890.5641.23
The validation dataset results, also shown in Figure 7, show that the SGD optimizer reaches the best performance in terms of Precision, with a value of 91.23%, although the Adam optimizer outperforms the SGD optimizer with a 94.74% value of Jaccard Index, 86.13% value of Dice Coefficient, 97.14% value of Recall, 95.01% of accuracy, and 16.24 of loss value. In the case of the Adadelta optimizer, the obtained results show that it is the worst one. Therefore, from these results we can affirm that the Adam optimizer has shown the best results on validation dataset, as it has outperformed on almost all parameters with respect to the SGD and Adadelta optimizers.
Figure 7

Analysis of confusion matrix parameters on different optimizers.

Figure 7 shows the analysis of the confusion matrix parameters on the Adam, Adadelta, and SGD optimizers using a validation dataset. From this figure, it can be seen that the Adam optimizer is performing best on almost all the parameters, such as Jaccard Index, Dice Coefficient, Precision, Recall, Accuracy, and Loss. The value of loss is much lower in the case of the Adam optimizer in comparison to the SGD and Adadelta optimizers.

3.2. Result Analysis Based on Different Optimizers

From Section 3.1, it is seen that the Adam optimizer has outperformed in comparison to the SGD and Adadelta optimizers with a batch size of 18. Therefore, in this section, the results are calculated using the Adam optimizer on different batch sizes. However, it is possible that the Adadelta and SGD optimizers may provide better results on different combinations of batch size and epochs. In future, these two optimizers can be analyzed for different batch size and epoch combinations. Here, the values of batch sizes used for analyzing the Adam optimizer are 8, 18, and 32 on 100 epochs.

3.2.1. Analysis of Training Loss and Accuracy

The results are taken using different batch sizes with the Adam optimizer on 100 epochs. Figure 8 shows the curves of training loss and training accuracy, and from the curves it can be concluded that the value of accuracy increases with the number of epochs, and the loss value decreases. The color red shows the training loss, and the color blue shows the training accuracy.
Figure 8

Analysis of training loss and training accuracy: (a) Training Loss on batch size 8; (b) Training Accuracy on batch size 8; (c) Training Loss on batch size 18; (d) Training Accuracy on batch size 18; (e) Training Loss on batch size 32; (f) Training Accuracy on batch size 32.

Figure 8a,c,e shows the training loss on batch sizes 8, 18, and 32, and the loss value is 0.5. Figure 8b,d shows the training accuracy on batch sizes 8 and 18, and the value of accuracy is approximately greater than 0.975. Figure 8f shows the training accuracy, and the value of accuracy is only 0.95 with a batch size of 32.

3.2.2. Analysis of Training Loss and Accuracy

Figure 9 shows the segmented images using the Adam optimizer with 100 epochs and a batch size of 8, 18, and 32. Figure 9a,c shows the ground truth masks of original images 1 and 2, and Figure 9b,d shows the original images. Figure 9e,g shows the predicted masks of original images 1 and 2 on batch size 8, whereas Figure 9f,h shows the segmented outputs of original images 1 and 2 on batch size 8. Similarly, Figure 9i–p shows the predicted masks and segmented outputs on batch sizes 18 and 32, respectively. From the visual analysis of the figures, it can be seen that batch sizes 8 and 18 show almost similar results with the Adam optimizer and 100 epochs, whereas batch size 32 does not perform well, since it is not extracting only the lesion part but also the outer part. Therefore, batch size 32 cannot be recommended for skin lesion segmentation. To see the best performing batch size between 8 and 18, the analysis of these two batch sizes, the confusion matrix parameters are evaluated in Section 3.2.3.
Figure 9

Images segmented on the Adam Optimizer, 100 epochs and different batch sizes: (a) Ground truth Mask of Original Image 1; (b) Original Image 1; (c) Ground truth Mask of Original Image 2; (d) Original Image 2; (e) Predicted Mask of Image 1 on batch size 8; (f) Segmented Output of Image 1 on batch size; (g) Predicted Mask of Image 2 on batch size 8; (h) Segmented Output of Image 2 on batch size 8; (i) Predicted Mask of Image 1 on batch size 18; (j) Segmented Output of Image 1 on batch size 18; (k) Predicted Mask of Image 2 on batch size 18; (l) Segmented Output of Image 2 on batch size 18; (m) Predicted Mask of Image 1 on batch size 32; (n) Segmented Output of Image 1 on batch size 32; (o) Predicted Mask of Image 2 on batch size 32; (p) Segmented Output of Image 2 on batch size 32.

3.2.3. Analysis of Confusion Matrix Parameters

In Section 3.2.2, a visual analysis of segmented images is done on different batch sizes, from which batch size 8 and 18 have shown the best results. Now, to see the best performing batch size, the confusion matrix parameters are analyzed. Table 3 shows the analysis of the U-Net model architecture using batch sizes 8, 18, and 32.
Table 3

Analysis of Different Batch sizes using the Adam Optimizer.

Training Dataset
Batch SizeJaccard Index (%)Dice Coefficient (%)Precision (%)Recall (%)Accuracy (%)Loss
897.6690.3797.1095.7897.827.90
1896.4288.3292.1598.5096.8811.31
3294.7980.8792.9396.0896.4517.02
Testing Dataset
Jaccard Index (%) Dice Coefficient (%) Precision (%) Recall (%) Accuracy (%) Loss
895.7287.2992.0494.1295.7712.54
1893.8384.8685.8996.9394.0419.19
3292.9278.3789.1993.2394.3421.41
Validation Dataset
Jaccard Index (%) Dice Coefficient (%) Precision (%) Recall (%) Accuracy (%) Loss
895.6887.4993.4292.7295.5113.72
1894.7486.1388.3097.1495.0116.24
3293.9279.7892.1393.2495.3019.19
In the case of the validation dataset, as also shown in Figure 10, the batch size of 18 has performed best on Recall with a value of 97.14%, although batch size 8 has outperformed and shown a 95.68% value of Jaccard Index, 87.49% value of Dice Coefficient, 93.42% value of Precision, 95.51% value of Accuracy, and a lower loss value, of 13.72. In the case of batch size 32, as already observed with the visual analysis, the performance is lower with respect to the other batch sizes, showing a loss of 19.19. Therefore, from these results it can be seen that batch size 8 has shown the best results on the validation dataset.
Figure 10

Analysis of confusion matrix parameters on different batch sizes.

Figure 10 shows the analysis of confusion matrix parameters on batch sizes 8, 18, and 32. From the figure it can be seen that batch size 8 is performing best on almost all the parameters, such as Jaccard Index, Dice Coefficient, Precision, Recall, Accuracy, and Loss. The value of loss is much lower in the case of batch size 8 in comparison to batch sizes 18 and 32.

3.3. Result Analysis Based on Different Epochs with the Adam Optimizer and Batch Size 8

From Section 3.2, it was seen that batch size 8 has outperformed in comparison to batch sizes 18 and 32 for the Adam optimizer. Therefore, in this section, the results are calculated using batch size 8 with different epochs. However, it is possible that batch sizes 18 and 32 may provide better results on different combinations of epochs. In future, these two batch sizes can be further analyzed with different epochs. Here, the value of epochs used for analyzing the Adam optimizer with batch size 8 are 25, 50, 75, and 100.

3.3.1. Analysis of Confusion Matrix Parameters

The results are taken using the Adam optimizer on batch size 8 with 25, 50, 75, and 100 epochs. Figure 11 shows the curves of training loss and training accuracy, and from the curves it is concluded that the value of accuracy increases with the number of epochs, and the loss value is decreases.
Figure 11

Analysis of training loss and training accuracy: (a) Training Loss on 25 epochs; (b) Training Accuracy on 25 epochs; (c) Training Loss on 50 epochs; (d) Training Accuracy on 50 epochs; (e) Training Loss on 75 epochs; (f) Training Accuracy on 75 epochs; (g) Training Loss on 100 epochs; (h) Training Accuracy on 100 epochs.

Figure 11a,c,e,g shows the training loss with 25, 50, 75, and 100 epochs, and the loss value is 0.5; Figure 11b shows the training accuracy on 25 epochs, and the value of accuracy is approximately greater than 0.94. Figure 11f,h shows the training accuracy, and the value of accuracy is only 0.975 on 75 and 100 epochs.

3.3.2. Visual Analysis of Segmented Images

Figure 12 shows the segmented images using the Adam optimizer and batch size 8 on different epochs. Figure 12a,c shows the ground truth masks of the original images 1 and 2, and Figure 12b,d shows the original images. Figure 12e,g shows the predicted masks of original images 1 and 2 on 25 epochs, whereas Figure 12f,h shows the segmented outputs of original images 1 and 2 on 25 epochs. Similarly, Figure 12i–t shows the predicted masks and segmented outputs on 50, 75, and 100 epochs, respectively.
Figure 12

Images segmented on the Adam Optimizer, batch size 8 and different epochs: (a) Ground truth Mask of Original Image 1; (b) Original Image 1; (c) Ground truth Mask of Original Image 2; (d) Original Image 2; (e) Predicted Mask of Image 1 on 25 epochs; (f) Segmented Output of Image 1 on 25 epochs; (g) Predicted Mask of Image 2 on 25 epochs; (h) Segmented Output of Image 2 on 25 epochs; (i) Predicted Mask of Image 1 on 50 epochs; (j) Segmented Output of Image 1 on 50 epochs; (k) Predicted Mask of Image 2 on 50 epochs; (l) Segmented Output of Image 2 on 50 epochs; (m) Predicted Mask of Image 1 on 75 epochs; (n) Segmented Output of Image 1 on 75 epochs; (o) Predicted Mask of Image 2 on 75 epochs; (p) Segmented Output of Image 2 on 75 epochs; (q) Predicted Mask of Image 1 on 100 epochs; (r) Segmented Output of Image 1 on 100 epochs; (s) Predicted Mask of Image 2 on 100 epochs; (t) Segmented Output of Image 2 on 100 epochs.

From the visual analysis of these figures, it can be seen that 25, 50, and 75 epochs show almost similar results on the Adam optimizer and batch size 8, whereas 100 epochs do not show good results. To see the best performing epochs between 25, 50, and 75, an analysis of these two epochs is done in Section 3.3.3 using confusion matrix parameters.

3.3.3. Analysis of Confusion Matrix Parameters

In Section 3.3.2, a visual analysis of segmented images is done on different epoch values, of 25, 50, 75, and 100. Now, to see the best performing epochs, the confusion matrix parameters are analyzed. Table 4 shows the analysis of the U-Net model architecture using 25, 50, 75, and 100 epochs size and a batch size of 8 with the Adam optimizer.
Table 4

Analysis of different epochs using the Adam optimizer and batch size 8.

Training Dataset
EpochsJaccard Index (%)Dice Coefficient (%)Precision (%)Recall (%)Accuracy (%)Loss
2588.6973.7281.7293.6991.5827.71
5093.51 79.81 98.74 81.03 93.62 18.99
7597.6690.7995.9596.8997.797.79
10059.97 53.07 37.62 96.75 47.37 164.86
Testing Dataset
Jaccard Index (%) Dice Coefficient (%) Precision (%) Recall (%) Accuracy (%) Loss
2589.7272.9580.0594.5891.6427.60
5093.10 78.97 96.55 81.10 93.35 19.44
7595.5787.4190.6295.2395.4713.78
10057.38 50.65 35.46 96.86 43.25 181.64
Validation Dataset
Jaccard Index (%) Dice Coefficient (%) Precision (%) Recall (%) Accuracy (%) Loss
2589.5673.9081.9692.3491.0028.17
5092.10 77.69 97.31 77.58 91.72 23.37
7596.3589.0193.5694.9196.2711.56
10059.78 53.26 37.58 96.73 47.15 165.86
From Table 4, in the case of the validation dataset, it can be seen that on 25 epochs the value of loss is 28.17, which is very high, followed by a loss value of 23.37 on 50 epochs, whereas on 75 epochs the value of loss becomes lower, i.e., 11.56. Moreover, the values of the Jaccard Index, Dice Coefficient, and Accuracy are increased. Therefore, it can be seen that during the training of the model, there was underfitting on 25 and 50 epochs, due to which the performance parameters values are not good. But at the epoch value of 75, the model is properly trained, so the parameters’ values are also improved. If the model is further trained up to 100 epochs, then the loss value is increased to 165.86. Hence, it can be identified that the proposed model performs best on 75 epochs. Figure 13 shows the analysis of confusion matrix parameters on 25, 50, 75, and 100 epochs. The results are obtained on the Jaccard Index, Dice Coefficient, Precision, Recall, Accuracy, and Loss. The best value of accuracy is obtained on 75 epochs with a much lower loss.
Figure 13

Analysis of confusion matrix parameters on different epochs.

3.4. Comparison with State-of-the-Art Techniques

A comparison of the suggested scheme with other current state-of-the-art methods using dermoscopy images has been performed in terms of both the Jaccard Coefficient and accuracy. Table 5 provides a breakdown of both class-level predictions. This result analysis shows that the proposed framework achieves a superior overall accuracy compared to the state-of-the-art approaches. Jaccard coefficient and accuracy differed from one study to the next, since they employed different datasets (ISBI-2016, ISBI-2017, and PH2). According to Yuan et al. [10,11], the Jaccard Coefficient is 0.963 for the ISBI-2016 dataset and 0.78 for the ISBI-2017 dataset when employing convolutional neural networks.
Table 5

Comparison of the Proposed Model with State-of-the-Art Techniques.

RefTechnique UsedDatasetPerformance Parameters
Yuan et al. [10]19-layer Deep Convolution NetworkISBI-2016Jaccard Coefficient = 0.963
PH2
Yuan et al. [11] Convolutional-Deconvolutional neural NetworkISBI-2017Jaccard Coefficient = 0.784
Hang Li et al. [28]Dense Deconvolutional NetworkISBI-2016Jaccard Coefficient = 0.870
ISBI-2017Jaccard Coefficient = 0.765
Yu et al. [29]Convolution NetworkISBI-2016Accuracy = 0.8654
ISBI-2017
Khan et al. [30]Convolution NetworkISICAccuracy = 0.968
PH2Accuracy = 0.921
Proposed Model Modified U-Net PH2 Jaccard Coefficient = 0.976
Architecture Accuracy = 0.977

4. Conclusions and Future Scope

Since medical image analysis is one of the challenging tasks which requires various computational techniques in the hierarchy of imaging applications, different types of analysis techniques, including image pre-processing, classification, segmentation, compression, and security, must be taken into account. In the literature, various authors have worked on the segmentation of skin lesions. This study proposed the modified U-Net model architecture for the segmentation of skin lesion in dermoscopy image so that an accurate classification of skin disease can be performed. The dermoscopy images are taken from the PH2 dataset with 200 images. The proposed model has been analyzed with different batch sizes, of 8, 18, and 32, using the Adam, Adadelta, and SGD optimizers and 25, 50, 75, and 100 epochs. The proposed model works best with a batch size of 8, the Adam optimizer, and 75 epochs, having an accuracy of 96.27% and a Jaccard Index of 96.35%. Its Dice Coefficient is coming out as 89.01%. Hence, there is still scope for improving the Dice Coefficient and the Precision of the modified U-Net architecture model. Moreover, in future, segmented images can be used for classification purposes to improve classification accuracy.
  21 in total

1.  Border detection in dermoscopy images using hybrid thresholding on optimized color channels.

Authors:  Rahil Garnavi; Mohammad Aldeen; M Emre Celebi; George Varigos; Sue Finch
Journal:  Comput Med Imaging Graph       Date:  2010-09-15       Impact factor: 4.790

2.  Combination of features from skin pattern and ABCD analysis for lesion classification.

Authors:  Zhishun She; Y Liu; A Damatoa
Journal:  Skin Res Technol       Date:  2007-02       Impact factor: 2.365

3.  A Mutual Bootstrapping Model for Automated Skin Lesion Segmentation and Classification.

Authors:  Yutong Xie; Jianpeng Zhang; Yong Xia; Chunhua Shen
Journal:  IEEE Trans Med Imaging       Date:  2020-02-10       Impact factor: 10.048

4.  Melanoma Recognition in Dermoscopy Images via Aggregated Deep Convolutional Features.

Authors:  Zhen Yu; Xudong Jiang; Feng Zhou; Jing Qin; Dong Ni; Siping Chen; Baiying Lei; Tianfu Wang
Journal:  IEEE Trans Biomed Eng       Date:  2018-08-20       Impact factor: 4.538

5.  Multiple skin lesions diagnostics via integrated deep convolutional networks for segmentation and classification.

Authors:  Mohammed A Al-Masni; Dong-Hyun Kim; Tae-Seong Kim
Journal:  Comput Methods Programs Biomed       Date:  2020-01-23       Impact factor: 5.428

6.  MultiResUNet : Rethinking the U-Net architecture for multimodal biomedical image segmentation.

Authors:  Nabil Ibtehaz; M Sohel Rahman
Journal:  Neural Netw       Date:  2019-09-04

7.  Automatic Skin Lesion Segmentation Using Deep Fully Convolutional Networks With Jaccard Distance.

Authors:  Yading Yuan; Ming Chao; Yeh-Chi Lo
Journal:  IEEE Trans Med Imaging       Date:  2017-04-18       Impact factor: 10.048

8.  Automatic skin lesion classification based on mid-level feature learning.

Authors:  Lina Liu; Lichao Mou; Xiao Xiang Zhu; Mrinal Mandal
Journal:  Comput Med Imaging Graph       Date:  2020-07-21       Impact factor: 4.790

9.  Evaluating the Usability of mHealth Applications on Type 2 Diabetes Mellitus Using Various MCDM Methods.

Authors:  Kamaldeep Gupta; Sharmistha Roy; Ramesh Chandra Poonia; Soumya Ranjan Nayak; Raghvendra Kumar; Khalid J Alzahrani; Mrim M Alnfiai; Fahd N Al-Wesabi
Journal:  Healthcare (Basel)       Date:  2021-12-21
View more
  3 in total

1.  A Secure Framework toward IoMT-Assisted Data Collection, Modeling, and Classification for Intelligent Dermatology Healthcare Services.

Authors:  Md Khairul Islam; Chetna Kaushal; Md Al Amin; Abeer D Algarni; Nazik Alturki; Naglaa F Soliman; Romany F Mansour
Journal:  Contrast Media Mol Imaging       Date:  2022-06-29       Impact factor: 3.009

2.  Content-Based Medical Image Retrieval System for Skin Melanoma Diagnosis Based on Optimized Pair-Wise Comparison Approach.

Authors:  Narendra Kumar Rout; Mitul Kumar Ahirwal; Mithilesh Atulkar
Journal:  J Digit Imaging       Date:  2022-10-17       Impact factor: 4.903

3.  Analysis of Vessel Segmentation Based on Various Enhancement Techniques for Improvement of Vessel Intensity Profile.

Authors:  Sonali Dash; Sahil Verma; SeongKi Kim; Jana Shafi; Muhammad Fazal Ijaz
Journal:  Comput Intell Neurosci       Date:  2022-06-28
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.