Literature DB >> 35722411

Development and validation of the interpretability analysis system based on deep learning model for smart image follow-up of nail pigmentation.

Yanqing Chen1,2, Haofan Liu3, Zhaoying Liu1, Yang Xie4, Yingxue Yao3, Xiaofen Xing3, Han Ma1,2.   

Abstract

Background: Nail pigmentation can be a clinical manifestation of malignant melanoma and a variety of benign diseases. Nail matrix biopsy for pathologic examination, the gold standard for diagnosis of subungual melanoma, is a painful procedure and may result in nail damage. Therefore, there is a great need for non-invasive methods and long-term follow-up for nail pigmentation. The objective of this study is to establish an intelligent precursor system to provide convenient monitoring for nail pigmentation, early warning subungual melanoma, and reduce nail biopsies to the minimum necessary.
Methods: Dermoscopic images of nail lesions were obtained from outpatients between 2019 and 2020. The images were divided into the training set and the test set using k-fold cross validation at a ratio of 10:1. The deep learning model is developed based on the Pytorch framework. The model structure is optimized using the image segmentation model U-Net. An image segmentation module analyzed the contours of the whole nail plate and pigmented area according to the boundary features of the input images and a rule calculation module used the output information of the segmentation model to automatically analyze specific indicators referring to the "ABCDEF" rule. The model's results were compared with those of clinical experts.
Results: From 550 dermoscopic images of nail lesions obtained, 500 were selected randomly as the training set, and the remaining 50 as the test set. Our image segmentation module realized automatic segmentation of the pigmented area and the whole nail plate with dice coefficient to be 0.8711 and 0.9652, respectively. Five qualitative indicators were selected in the interpretability analysis system and the models showed a certain degree of consistency with the evaluation by clinical experts, i.e., R2 for area ratio vs. breadth score was 0.8179 (P<0.001), for mean pixel value vs. pigment score was 0.7149 (P<0.001), for evenness vs. pigment score was 0.5247 (P<0.001). Conclusions: The proposed system made accurate segmentation of the nail plate and pigmented area and achieved medically interpretable index analysis. It is potentially a primer of an intelligent follow-up system that will enable convenient and spatially unaffected management and monitoring of nail pigmentation. It may provide clinicians with understandable auxiliary information for diagnosis. 2022 Annals of Translational Medicine. All rights reserved.

Entities:  

Keywords:  Deep learning; dermoscopy; interpretability; melanoma; nail pigmentation

Year:  2022        PMID: 35722411      PMCID: PMC9201122          DOI: 10.21037/atm-22-1714

Source DB:  PubMed          Journal:  Ann Transl Med        ISSN: 2305-5839


Introduction

Brown or black pigmentation within the fingernails or the toenails is a common presenting problem in dermatology. Common causes of nail pigmentation are benign disorders such as melanocytic nevus, subungual hemorrhage, onychomycosis, exogenous staining, and systemic diseases. However, it may also be a sign of a life-threatening malignant disease: nail apparatus melanoma. Nail apparatus melanoma accounts for about 2% of all melanomas, and up to 20% of melanomas in dark-skinned and Asian populations (1,2). Early diagnosis and treatment greatly improve the prognosis, but unfortunately, melanoma of the nail unit is usually diagnosed at a later stage, which leads to poor prognosis. The gold standard for the diagnosis of nail apparatus melanoma is histopathological examination of a biopsy specimen of the nail matrix, which is usually a painful procedure and may result in nail dystrophy, permanent deformity, and paresthesia (3). Therefore, it is often rejected by patients. Non-invasive methods to assist in the diagnosis and follow-up of nail apparatus melanoma can balance missed diagnosis and excessive nail biopsies. A clinical analysis system, named the “ABCDEF” rule, was developed to evaluate the risk of suspected cases of nail apparatus melanoma (4,5). “A” stands for age, African Americans, Asians, and native Americans; “B “stands for brown to black band with breadth of 3 mm or more and variegated borders; “C” stands for change in the nail band; “D” stands for the digit involved; “E” stands for extension of the pigment onto the nailfold; and “F” stands for family or personal history (4,5). To further establish the diagnosis, dermoscopy, a non-invasive procedure obtaining images by a dermoscope, has been used to evaluate the characteristics of nail lesions, and helps to dynamically observe changes in the pigmentation of the nail (4). Remote intelligent system is a promising tool for improving patient compliance with disease surveillance. Artificial intelligence in medical imaging has made remarkable progress in recent years and has the potential to greatly affect clinical disease management (6,7). However, deep learning models usually work similar to a black box, resulting in a lack of interpretability. Most studies of melanoma based on deep learning algorithms are classification, segmentation, and detection (8,9). Compared with traditional image algorithms, such end-to-end convolutional neural network (CNN) models usually work without feature engineering with the support of high-quality data (8,9). The poor interpretability of CNN models in clinical practice is risky and makes it difficult to gain the trust of clinicians and patients (10). In order to improve this situation, we developed and validated an indicator analysis system based on a deep learning image segmentation model of pigmentations in the nail. The proposed analysis system first identifies pigmented lesions on nails, and then extract relevant indicators with reference to the established “ABCDEF” rule to provide interpretable data. The aim of the systems is to assist in dynamic monitoring of nail pigmentation, early warning of subungual melanoma, and prompting the best timing from non-invasive follow-up to necessary invasive examination in an interpretable manner. We present the following article in accordance with the TRIPOD reporting checklist (available at https://atm.amegroups.com/article/view/10.21037/atm-22-1714/rc).

Methods

Image data set

We included patients who presented to the outpatient for abnormal pigmentation within the nail at the Department of Dermatology in Fifth Affiliated Hospital of Sun Yat-sen University in China between March 1, 2019 and October 31, 2020. The diagnostic categories included: nail matrix nevus, malignant melanoma, subungual hemorrhage, onychomycosis, subungual glomus tumor, etc. Dermoscopic images of the nail lesions were taken with the same dermascope. The average image size was ≈1,500×2,400 dpi. The dermoscopic images were randomly divided into the training set and the test set for the proposed deep learning segmentation model using k-fold cross validation at a ratio of 10:1. To develop the deep learning model, the dermoscopic images were input as the source samples. Each image was annotated by the image labeling tool Labelme to mark the contours of the nail area and pigmented spots or lines respectively. A set of binary gray images was generated as the training label for the deep learning segmentation model. In order to reduce the over-fitting problem during the training process of the segmentation model and enhance the adaptability of the model in clinical scenarios, we performed data enhancement operations such as rotation, translation, cropping and noise addition on the training set. The test set was used to evaluate the effect of the segmentation model. The study was conducted in accordance with the Declaration of Helsinki (as revised in 2013). The study was approved by the Ethics Committee of the Fifth Affiliated Hospital of Sun Yat-sen University (No. [2021](K225-1)) and all patients gave informed consent.

Automatic analysis system

Our nail pigmentation automatic segmentation system consisted of two modules: image segmentation and rule calculation module. A deep learning image segmentation model was used in the automatic image segmentation module because it can accurately recognize the contours of the whole nail plate and pigmented line area, respectively, according to the boundary features of the input images. The rule calculation module is an intelligent analysis module connected with the segmentation module. It makes use of the output information of the segmentation model to automatically analyze specific indicators of the dermoscopic nail images. The system was developed using the Python open-source deep learning framework PyTorch. Two different U-Net models were used as the image segmentation model. For each dermoscopic nail image, the two U-Net models output the contours of the whole nail and dark spots or black lines areas respectively. Referring to “ABCDEF” rule, breadth of nail band, border, digit involvement, and pigment extension of the nail band were features which can be reflected through images. Four individual experienced dermatologists scored the breath, pigment, border, extension (on the nailfold) and the general risk (to be malignant) for the nail images. Each item has a 10-step score, ranging from 0 to 9, whereby the score 9 reflects the most consistent with malignant. Using the results of the image segmentation model, the rule analysis module analyzed five qualitative indicators which reflect the above features. These indicators include area ratio, mean pixel value, evenness, irregularity, and skin invasion. The area ratio indicator is the ratio of the number of pixels in dark spots or black lines and the whole nail area. Mean pixel value indicator was obtained by calculating the average pixel value ratio of the dark spots or black lines relative to the whole nail area. Evenness indicator was represented by the pixel value standard deviation. The irregularity indicator, which means the irregularity of the shape of dark spots and black lines, was evaluated it by computing the Hu moments gap between the target area contour and its minimum circumscribed rectangle. The skin invasion indicator was calculated as the percentage of the non-intersecting part of nail area and black line area to the black line area. We then analyzed the consistency of these indicators with features of the nail pigmentation evaluated by dermatologists.

Statistical analysis

To evaluate the performance of the image segmentation model, pixel accuracy (PA), Jaccard index, and the dice coefficient (DC) were used. The U-Net model predicts the classification probability of the pixels of the input image, which is divided into foreground (target area) and background, and outputs a pixel probability matrix with the same size as the original image. Each value in the matrix corresponds to the classification probability of the pixel at the same position in the source image. The value is between 0 and 1. Pixels with a probability value >0.5 are classified as foreground according to prior knowledge. Linear regression analysis was used to study the relationship between the representative indicators and clinical score evaluated by experts. R square (R2, coefficient of determination) was used to determine the strength of the correlation between the indicators and the clinical features assessed by experts. R2 value ranges from 0 to 1, with over 0.7 considered a good fit, between 0.5–0.7 was considered a moderate fit for the given model. The statistical analyses were performed by using GraphPad Prism 6 and P<0.05 was considered significant.

Results

Accurate segmentation of nail images in the proposed automatic segmentation system

A general flowchart of the indicator analysis system for images of nail pigmentation is presented in . Among the data set of 550 dermoscopic images collected from outpatients with nail pigmentation, 500 images were classified randomly as the training set and the remaining 50 comprised the test set. Accurate segmentation ensures the accuracy and reliability of the rule calculation module. With the image labeling tool, the contours of the nail area () and pigmented spots or lines () were accurately marked. is a representative example of the results of the segmentation of the nail images, including segmentation of a pigmented line area () and the whole nail plate (). Applying the trained image segmentation model to the test set, accuracy of the model was evaluated using PA, Jaccard index, and DC (). The results showed that the proposed nail image segmentation model had a good or acceptable segmentation effect on the target area. Notably, the DC reached 96.52% and PA reached 97.85% in the segmentation of the nail area. In addition, the segmentation of the nail area achieved higher PA, Jaccard index, and DC than the pigmented spot or line area. In summary, the results suggested that accurate segmentation can be achieved using the proposed model.
Figure 1

Flowchart of the indicator analysis system for images of nail pigmentation.

Figure 2

Segmentation label and results of the algorithms. (A) Nail area segmentation; (B) pigmented spot or line area segmentation. FCN, fully convolutional network.

Figure 3

Segmentation results of the nail images in the proposed model. (A) Pigmented spot or line area segmentation; (B) Nail area segmentation. The right-hand image is the original segmentation label in each sample pair. Specifically, the left-hand image is the visualization of the segmentation result, and red area in the left-hand image is the segmentation result outputted by the model. The result shows that the nail image segmentation model was robustness and showed significant segmentation of the target area.

Table 1

Accuracy evaluation of the segmentation of the nail area and the pigmented spot or line area

Target areaPAJaccardDC
Nail area0.97850.93320.9652
Pigmented spot or line area0.97530.78420.8711

PA, pixel accuracy; DC, dice coefficient.

Flowchart of the indicator analysis system for images of nail pigmentation. Segmentation label and results of the algorithms. (A) Nail area segmentation; (B) pigmented spot or line area segmentation. FCN, fully convolutional network. Segmentation results of the nail images in the proposed model. (A) Pigmented spot or line area segmentation; (B) Nail area segmentation. The right-hand image is the original segmentation label in each sample pair. Specifically, the left-hand image is the visualization of the segmentation result, and red area in the left-hand image is the segmentation result outputted by the model. The result shows that the nail image segmentation model was robustness and showed significant segmentation of the target area. PA, pixel accuracy; DC, dice coefficient.

Interpretability of the quantitative indicators of the proposed model

To achieve clinical interpretability, we devised some digitally comparable features that can be extracted according to the “ABCDEF” rule, especially “B” and “E”, which are closely related to the morphological features of the lesions. Five important indicators were selected and named the area ratio, mean pixel value, evenness, irregularity, and skin invasion respectively (). To analyze the clinical significance of these indicators, five representative nail pigmentation images were presented for index analysis (). The breadth, border, and nailfold extension of the pigmented lesions, which are clinical aspects represented by “B” and “E” in the “ABCDEF” rule, were scored by experienced dermatologists and quantitatively reflected in these indicators (). Linear regression analysis showed that the numerical value output by the computer had a similar change trend as the evaluation value of clinical experts. Further, we found that the indicator area ratio and mean pixel value were of good consistency with the judgments of clinical experts, with the R2 for area ratio vs. breadth score to be 0.8179 (P<0.001), for mean pixel value vs. pigment score to be 0.7149 (P<0.001) (). Acceptable consistency was observed between evenness and pigment score (R2=0.5247, P<0.001) (). Therefore, from the comparison of the sample images with quantitative indicators, we concluded that our intelligent analysis system has the potential to develop into an interpretable intelligent nail pigmentation follow-up system.
Figure 4

Radar chart showing the features of the 5 qualitative indicators indexed by the proposed model.

Table 2

Expert scores and advice for the characteristics of five representative nail pigmentation images

Image No.Breadth scoreBorder scorePigment scoreExtension scoreMalignant/benignMedical advice
197.598.59Biopsy as soon as possible
229817Biopsy
349899Biopsy as soon as possible
413602Follow up
50.50201No intervention required
Table 3

Index analysis results for representative nail images with pigmented line

ItemImage No.
12345
Representative image
Area ratio0.910.260.500.110.10
Mean pixel value35.8670.4354.8141.45140.05
Evenness18.2724.2823.3120.909.33
Irregularity2.012.101.070.880.03
Skin invasionTrueFalseTrueFalseFalse
Figure 5

Regression analysis of the representative qualitative indicators and clinical expert assessment. (A) Area ratio vs. breadth score (P<0.001); (B) mean pixel value vs. pigment score (P<0.001); (C) evenness vs. pigment score (P<0.001); and (D) irregularity vs. border score (P=0.0014).

Radar chart showing the features of the 5 qualitative indicators indexed by the proposed model. Regression analysis of the representative qualitative indicators and clinical expert assessment. (A) Area ratio vs. breadth score (P<0.001); (B) mean pixel value vs. pigment score (P<0.001); (C) evenness vs. pigment score (P<0.001); and (D) irregularity vs. border score (P=0.0014).

Discussion

Deep learning models have been applied to various scenarios of medical image recognition and classification (11-13). For example, Esteva et al. proposed a dermatologist-level classification model that can distinguish keratinocyte carcinomas and malignant melanomas for binary classification data and the performance of this system is comparable to that of clinical experts (11); Han et al. established a deep learning model of classification of clinical images of 12 skin diseases (12); and Yu et al. used a CNN with dermoscopic images of acral melanoma and benign nevi (13). In view of the different structures of nail and skin, the current artificial intelligence diagnosis system is currently insufficient for nail images specifically. As a risk evaluation for malignant melanoma, the “ABCDEF” rule for nails is also different from that for other body parts. Furthermore, it is increasingly recognized that the so-called “black box” working method of deep learning models is an important obstacle to their widespread use in clinical practice (10,14). Having unexplainable outcomes will always make it difficult to gain the confidence of clinicians and patients on health issues (10,14). Long-term clinical follow-up is an important aspect of early diagnosis of nail apparatus melanoma, and for reducing the number of nail biopsies to the minimum necessary. Therefore, an interpretability analysis system based on a deep learning model of nail images will have positive clinical benefit if it can provide reliable and interpretable information of the characteristics and risk of nail pigmentation. The proposed segmentation model showed good potential because of it performed well in segmentation of the nail images. An accurate segmentation model will first extract the target area and eliminate the interference of background shadows. Next, a connected rule-based module will give an intuitive analysis of the nail-specific “ABCDEF” rule as a clinical evaluation. For the proposed model in this study, some digitally comparable characteristics that were consistent with “B” and “E” of the “ABCDEF” rule were extracted. The model outcomes have good or acceptable consistency with the expert evaluation. Area ratio corresponded to breadth, mean pixel value and evenness corresponded to pigment, irregularity corresponded to border, and skin invasion corresponded to extension. With long-term follow-up comparing these characteristics at different times, the “C”, which means changes, would eventually be estimated with image series. Therefore, with the proposed system, we can perform the index analysis and features interpretation based on dermoscopic images and finally achieve dynamic, intelligent follow-up and warning of high risk. Compared with a simple classification model, our intelligent analysis system conforms to the actual clinical diagnosis process, and the specific indicator analysis is interpretable, which is more suitable for practical applications. Deep learning models usually have a mass of parameters that lead to slow calculation speed. We performed convolutional layer compression on the feature extraction module of the segmentation model, which saved computing resources and memory resources while maintaining the same segmentation performance. In order to make the intelligent analysis system more sensitive and robust for images of nail pigment lesions, the dataset has to be enlarged to further increase the accuracy and reliability of the segmentation and index analysis results. Our study has provided a prototype artificial intelligence model with good potential for clinical application. Further studies are needed to enrich the index types, optimize the index analysis formula, incorporate more modal information, and build a multimodal analysis system for nail pigmentation. As the first step, the analysis system based on a dermoscopic image database is still insufficient to adapt to mobile phone images. In the future, the improvement and application of this analysis system to images obtained by mobile phones will help to monitor diseases more conveniently, without being restricted by location and dermascopic equipment. In addition, the training of this model is based on the judgement of experts and therefore there may be bias. However, bias could be reduced in the future by collecting opinion from more experienced clinicians and referring to updated recognized standards. In addition, as the indicators in this study are interpretable and have similar logic to clinical practice, when bias occurs, the model can be corrected retrospectively. The article’s supplementary files as
  12 in total

1.  Explaining the Rationale of Deep Learning Glaucoma Decisions with Adversarial Examples.

Authors:  Jooyoung Chang; Jinho Lee; Ahnul Ha; Young Soo Han; Eunoo Bak; Seulggie Choi; Jae Moon Yun; Uk Kang; Il Hyung Shin; Joo Young Shin; Taehoon Ko; Ye Seul Bae; Baek-Lok Oh; Ki Ho Park; Sang Min Park
Journal:  Ophthalmology       Date:  2020-06-26       Impact factor: 12.079

2.  Site distribution of cutaneous melanoma in South Korea: a retrospective study at a single tertiary institution.

Authors:  Jun-Hwan Kim; Ji-Ho Park; Dong-Youn Lee
Journal:  Int J Dermatol       Date:  2014-10-14       Impact factor: 2.736

3.  Classification of the Clinical Images for Benign and Malignant Cutaneous Tumors Using a Deep Learning Algorithm.

Authors:  Seung Seog Han; Myoung Shin Kim; Woohyung Lim; Gyeong Hun Park; Ilwoo Park; Sung Eun Chang
Journal:  J Invest Dermatol       Date:  2018-02-08       Impact factor: 8.551

4.  Artificial intelligence in medical imaging.

Authors:  John C Gore
Journal:  Magn Reson Imaging       Date:  2019-12-16       Impact factor: 2.546

5.  Melanoma detection using adversarial training and deep transfer learning.

Authors:  Hasib Zunair; A Ben Hamza
Journal:  Phys Med Biol       Date:  2020-07-06       Impact factor: 3.609

Review 6.  Management of Nail Unit Melanoma.

Authors:  Junqian Zhang; Sook Jung Yun; Stacy L McMurray; Christopher J Miller
Journal:  Dermatol Clin       Date:  2021-02-11       Impact factor: 3.478

7.  Dermoscopic examination of nail pigmentation.

Authors:  Sandra Ronger; Sandrine Touzet; Claire Ligeron; Brigitte Balme; Anne Marie Viallard; Danièle Barrut; Cyrille Colin; Luc Thomas
Journal:  Arch Dermatol       Date:  2002-10

Review 8.  Subungual Melanoma.

Authors:  Travis W Littleton; Peter M Murray; Mark E Baratz
Journal:  Orthop Clin North Am       Date:  2019-07       Impact factor: 2.472

9.  Dermatologist-level classification of skin cancer with deep neural networks.

Authors:  Andre Esteva; Brett Kuprel; Roberto A Novoa; Justin Ko; Susan M Swetter; Helen M Blau; Sebastian Thrun
Journal:  Nature       Date:  2017-01-25       Impact factor: 49.962

Review 10.  Artificial intelligence in radiology.

Authors:  Ahmed Hosny; Chintan Parmar; John Quackenbush; Lawrence H Schwartz; Hugo J W L Aerts
Journal:  Nat Rev Cancer       Date:  2018-08       Impact factor: 60.716

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.