| Literature DB >> 32292713 |
Nils Gessert1,2, Maximilian Nielsen2,3, Mohsin Shaikh2,3, René Werner2,3, Alexander Schlaefer1,2.
Abstract
In this paper, we describe our method for the ISIC 2019 Skin Lesion Classification Challenge. The challenge comes with two tasks. For task 1, skin lesions have to be classified based on dermoscopic images. For task 2, dermoscopic images and additional patient meta data are used. Our deep learning-based method achieved first place for both tasks. The are several problems we address with our method. First, there is an unknown class in the test set which we cover with a data-driven approach. Second, there is a severe class imbalance that we address with loss balancing. Third, there are images with different resolutions which motivates two different cropping strategies and multi-crop evaluation. Last, there is patient meta data available which we incorporate with a dense neural network branch. • We address skin lesion classification with an ensemble of deep learning models including EfficientNets, SENet, and ResNeXt WSL, selected by a search strategy. • We rely on multiple model input resolutions and employ two cropping strategies for training. We counter severe class imbalance with a loss balancing approach. • We predict an additional, unknown class with a data-driven approach and we make use of patient meta data with an additional input branch.Entities:
Keywords: Convolutional neural networks; Deep Learning; Multi-class skin lesion classification
Year: 2020 PMID: 32292713 PMCID: PMC7150512 DOI: 10.1016/j.mex.2020.100864
Source DB: PubMed Journal: MethodsX ISSN: 2215-0161
Fig. 1Cropping strategy for dermoscopic images with a large, black area around the images.
Fig. 2General approach for combining dermoscopic image processing and meta data processing.
All cross-validation results for different configurations. We consider same-sized cropping (SS) and random-resize cropping (RR) and different model input resolutions. Values are given in percent as mean and standard deviation over all five CV folds. Ensemble average refers to averaging over all predictions from all models. Ensemble optimal refers to averaging over the models we found with our search strategy for the optimal subset of configurations. C = 8 refers to training with eight classes without the unknown class. T1 refers to Task 1 without meta data and T2 refers to Task 2 with meta data. ResNext WSL 1 and 2 refer to ResNeXt-101 WSL 32 × 8d and 32 × 16d, respectively [8].
| Configuration | Sensitivity T1 | Sensitivity T2 |
|---|---|---|
| SENet154 SS 224 × 224 | 67.2 ± 0.8 | 70.0 ± 0.8 |
| ResNext WSL 1 SS 224 × 224 | 65.9 ± 1.6 | 68.1 ± 1.3 |
| ResNext WSL 2 SS 224 × 224 | 65.3 ± 0.8 | 69.1 ± 1.5 |
| EN B0 SS 224 × 224 C = 8 | 66.7 ± 1.8 | 68.8 ± 1.5 |
| EN B0 SS 224 × 224 | 65.8 ± 1.7 | 67.4 ± 1.6 |
| EN B0 RR 224 × 224 | 67.0 ± 1.6 | 68.9 ± 1.7 |
| EN B1 SS 240 × 240 | 65.9 ± 1.6 | 68.2 ± 1.8 |
| EN B1 RR 240 × 240 | 66.8 ± 1.5 | 68.5 ± 1.8 |
| EN B2 SS 260 × 260 | 67.2 ± 1.4 | 69.0 ± 2.5 |
| EN B2 RR 260 × 260 | 67.6 ± 2.0 | 70.1 ± 2.0 |
| EN B3 SS 300 × 300 | 67.8 ± 2.0 | 68.5 ± 1.7 |
| EN B3 RR 300 × 300 | 67.0 ± 1.5 | 68.4 ± 1.5 |
| EN B4 SS 380 × 380 | 67.8 ± 1.1 | 68.5 ± 1.1 |
| EN B4 RR 380 × 380 | 68.1 ± 1.6 | 69.4 ± 2.2 |
| EN B5 SS 456 × 456 | 68.2 ± 0.9 | 68.7 ± 1.6 |
| EN B5 RR 456 × 456 | 68.0 ± 2.2 | 69.0 ± 1.6 |
| EN B6 SS 528 × 528 | 68.8 ± 0.7 | 69.0 ± 1.4 |
| Ensemble Average | 71.7 ± 1.7 | 73.4 ± 1.6 |
| Ensemble Optimal | 72.5 ± 1.7 | 74.2 ± 1.1 |
| Official Testset | 63.6 | 63.4 |
Results from the official test set of the ISIC 2019 Challenge for each class. We consider the AUC, the AUC for a sensitivity larger than 80% (AUC-S), the sensitivity and specificity. Note that the sensitivity given here is differently calculated than S. Values are given in percent.
| Class | Task1 | Task2 | ||||||
|---|---|---|---|---|---|---|---|---|
| AUC | AUC-S | Sens. | Spec. | AUC | AUC-S | Sens. | Spec. | |
| MEL | 0.928 | 0.849 | 0.594 | 0.962 | 0.931 | 0.849 | 0.545 | 0.976 |
| NV | 0.96 | 0.93 | 0.71 | 0.975 | 0.96 | 0.932 | 0.637 | 0.983 |
| BCC | 0.949 | 0.904 | 0.721 | 0.94 | 0.947 | 0.901 | 0.649 | 0.958 |
| AK | 0.914 | 0.824 | 0.484 | 0.965 | 0.919 | 0.841 | 0.46 | 0.966 |
| BKL | 0.904 | 0.805 | 0.394 | 0.985 | 0.908 | 0.821 | 0.324 | 0.991 |
| DF | 0.979 | 0.963 | 0.578 | 0.992 | 0.98 | 0.965 | 0.556 | 0.993 |
| VASC | 0.956 | 0.925 | 0.644 | 0.991 | 0.942 | 0.912 | 0.495 | 0.995 |
| SCC | 0.938 | 0.876 | 0.439 | 0.986 | 0.93 | 0.878 | 0.408 | 0.987 |
| UNK | 0.775 | 0.581 | 0.00283 | 0.999 | 0.612 | 0.253 | 0 | 0.999 |
Specifications Table
| Subject Area: | Computer Science |
| More specific subject area: | Deep learning and skin lesion classification |
| Method name: | Convolutional neural network |
| Name and reference of original method: | Not applicable – our method is based on multiple approaches which we cite and detail in the method description |
| Resource availability: | Public code: |