Literature DB >> 25941190

The feasibility of using manual segmentation in a multifeature computer-aided diagnosis system for classification of skin lesions: a retrospective comparative study.

Wen-Yu Chang1, Adam Huang2, Yin-Chun Chen3, Chi-Wei Lin4, John Tsai5, Chung-Kai Yang6, Yin-Tseng Huang7, Yi-Fan Wu8, Gwo-Shing Chen9.   

Abstract

OBJECTIVES: To investigate the feasibility of manual segmentation by users of different backgrounds in a previously developed multifeature computer-aided diagnosis (CADx) system to classify melanocytic and non-melanocytic skin lesions based on conventional digital photographic images.
METHODS: In total, 347 conventional photographs of melanocytic and non-melanocytic skin lesions were retrospectively reviewed, and manually segmented by two groups of physicians, dermatologists and general practitioners, as well as by an automated segmentation software program, JSEG. The performance of CADx based on inputs from these two groups of physicians and that of the JSEG program was compared using feature agreement analysis.
RESULTS: The estimated area under the receiver operating characteristic curve for classification of benign or malignant skin lesions based were comparable on individual segmentation by the gold standard (0.893, 95% CI 0.856 to 0.930), dermatologists (0.886, 95% CI 0.863 to 0.908), general practitioners (0.883, 95% CI 0.864 to 0.903) and JSEG (0.856, 95% CI 0.812 to 0.899). The agreement in the malignancy probability scores among the physicians was excellent (intraclass correlation coefficient: 0.91). By selecting an optimal cut-off value of malignancy probability score, the sensitivity and specificity were 80.07% and 81.47% for dermatologists and 79.90% and 80.20% for general practitioners.
CONCLUSIONS: This study suggests that manual segmentation by general practitioners is feasible in the described CADx system for classifying benign and malignant skin lesions. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://group.bmj.com/group/rights-licensing/permissions.

Entities:  

Mesh:

Year:  2015        PMID: 25941190      PMCID: PMC4420958          DOI: 10.1136/bmjopen-2015-007823

Source DB:  PubMed          Journal:  BMJ Open        ISSN: 2044-6055            Impact factor:   2.692


SKINCAD, an effective computer-aided diagnosis (CADx) system developed in our previous study, achieved performance similar to face-to-face clinical diagnosis by staff dermatologists at our institution. Regarding clinical application, evaluating the feasibility of subjective manual segmentation by users of different backgrounds is especially useful for images with complicated components, whereas automated segmentation methods sometimes fail. This study simulates a real clinical setting and groups, general practitioners (GPs) and dermatologists, have manually segmented a wide spectrum of skin lesions under generous inclusion criteria, which represents skin lesions encountered in daily practice, with each lesion given a definite histopathological diagnosis. Our result suggests that GP-determined borders performed as well as dermatologist-determined borders in CADx diagnosis by SKINCAD for classification of benign or malignant skin lesions. Through the in-depth evaluation of overlap index and feature agreement levels, our study indicates the possibility of direct onsite computation application for physicians other than dermatology specialists when assessing melanocytic and non-melanocytic skin lesions. This retrospective analysis was restricted to biopsied lesions performed in a single medical centre, as we used the histopathological reports as the gold standard. Further large-scale prospective study may be required in the future for broader application.

Introduction

Skin cancer is a common malignancy worldwide. The increasing cost of skin cancer management over the last decade constitutes a substantial health problem.1–3 With lower incidence rates of melanoma in Asians than in Caucasians, non-melanoma skin cancers, such as squamous cell carcinoma and basal cell carcinoma, contribute to significant morbidities as well, especially in the Asian population. In clinical practice, most physicians detect skin cancer by visual examination, which is highly dependent on experience and specialisation. Although the accuracy rate of clinicians can be improved with the support of dermoscopy when confronted with difficult-to-diagnose skin lesions, this approach relies on the specific training of a limited population of clinicians, mainly dermatological specialists who manage skin tumours. Previously, we developed an effective computer-aided diagnosis (CADx) system (SKINCAD), which classifies melanocytic and non-melanocytic skin lesions by utilising conventional digital macrophotographs. This system achieved performance similar to face-to-face clinical diagnosis by staff dermatologists at our institution.4 In the study, a dermatologist manually segmented the images for analysis. Regarding the concerns of subjectivity and consistency of the manually generated borders,5 6 many automatic border detection methods have been developed, such as the JSEG algorithm, contrast enhancement and clustering algorithms, with different evaluation metrics.6–11 However, many of these algorithms were developed to approximate the ground truth borders, which were also determined by dermatologists subjectively.7 8 10 12 13 Furthermore, because of the complexity of skin images, it is not usually possible for every lesion to be segmented automatically. In a study that compared three dermoscopic image analysis systems, approximately half of the skin lesions were not analysable by at least one of the three systems due to programming limitations, such as the inability to perform segmentation, and the operator had to adjust the computer-determined segmentation manually.14 We would expect that the segmentation task might be more challenging when clinical digital photographs are used. In our previous study, the use of the CADx system could be repeated with relative consistency by users without medical training.15 The purpose of this study was to investigate the feasibility and reliability of manual segmentation performed by medical practitioners of different backgrounds (general practitioners (GPs) or board-certified dermatologists), and to compare their performance with that of a commonly used autosegmentation algorithm, JSEG, in a multifeature, CADx system, SKINCAD. In particular, this study aimed to assess the potential use of this system by GPs without skin cancer training.

Materials and methods

Data acquisition

From January 2010 to December 2010, 2148 consecutive skin lesions were biopsied or excised by dermatologists for histological confirmation in the Department of Dermatology, Kaohsiung Medical University. A total of 13 908 digital photographs of these lesions were taken, prior to the biopsy and excision procedures, for recording purposes. We retrospectively reviewed all cases and excluded non-tumour specimens, lesions that had undergone previous surgical procedures and images that were misregistered or of poor quality (unfocused or motion blurred). The database consists of 1151 images from 295 patients (347 specimens). All of the lesions were examined at the clinic by board-certified staff dermatologists from our institute. A dermatologist (W-YC, with 11 years of experience) carefully reviewed these images and selected a representative close-up image for each lesion. Images with hair artefacts were not excluded because the preselection of data showed little influence based on a large data set in a previous report.16 Photographs were obtained with a 6.1-megapixel digital single-lens reflex camera (D70, Nikon Corporation, Tokyo, Japan) with an 18–50 mm F2.8 macrolens (Sigma Corporation, Fukushima, Japan). When obtaining close-up images, the target lesion was focused and positioned at the centre of the photographs, and the size was controlled so as not to exceed 50% of the area of the photograph.

CADx system

The in-house CADx system used in this study, SKINCAD, was developed and optimised in our previous report to include 16 effective features, with optimal operation parameters for an Asian database.4 In brief, it is a CADx system that applies various image processing algorithms to extract the shape, colour and texture features, not only for the skin lesion itself but also for the surrounding rectangular cropped area. The raw score assigned after the CADx evaluation is adjusted to be the probability of malignancy according to the previous database model calibrated using Platt's method.4 17

Manual segmentation study

Six months after the images were collected, four board-certified family physicians and three dermatologists (including W-YC) with an average experience of 10 years were asked to draw the skin lesion borders for all 347 digital images using the CADx system's graphical user interface (GUI), which was developed using MATLAB software (figure 1). When the physician started to use the software, a brief introduction was presented with 10 sample images demonstrating how to use the software to mark the borders and how to save the results. The software GUI showed one lesion at a time, and the reader could use the pan/zoom function as needed for observation. To simulate a real clinical setting, there were no preselection processes, specific instructions or feedback regarding the actual performance of their segmentation results. The physicians evaluated and marked the lesion border according to their own clinical experience, and they were allowed to use their own equipment for convenience, as long as they were confident enough to accurately define the tumour lesion area and differentiate it from normal skin. Two physicians chose to use a pen tablet (CTH-661, Wacom, Taiwan), and the others chose to use their own mouse as a pointing and border-drawing device.
Figure 1

The SKINCAD system graphical user interface developed using MATLAB software.

The SKINCAD system graphical user interface developed using MATLAB software.

Automated segmentation study

There are many image segmentation algorithms, and it is challenging to compare them. In our previous report, we evaluated a few software programs that are available online,8 18 19 and tested them by using a diverse data set of 769 images with 19 categories of histological diagnoses.15 Owing to the complexity of the clinical images, a non-medical individual performed the preselection of an area of interest prior to the automated segmentation phase of all methods. Among the methods we tested preliminarily, JSEG algorithm18 performed the best with respect to all three criteria, those being, good colour capability, short computational time and consistent accuracy. JSEG was chosen as our automated segmentation algorithm because of its flexibility and good performance in a variety of domains such as natural scenery,18 colonoscopy images,20 tongue images21 and skin lesion images.13

Overlapping and variability

The average or comparable extraction results obtained by multiple dermatologists can be considered more reliable.12 In this study, we defined the area extracted by two or more dermatologists as the gold standard of the ground-truth tumour area. To assess the reliability of the performance of CADx using manual segmentation and the variability introduced by each user, we computed the following: Intraclass correlation coefficient (ICC) to assess inter-rater reliability of the malignancy probability scores among multiple raters.22 Area overlap percentage or Jaccard index,23 as defined in Eq. (1), between each GP and the gold standardwhere SGP and SGold standard denote the lesion areas determined by the borders drawn by GP and the gold standard, respectively, and and are the intersection and union, respectively. Lesion inclusion rate: the percentage of the ground-truth lesion included in the cropped image area (rectangular image generated by software to enclose the entire manual segmentation) as defined in Eq. (2) The above definitions are illustrated in figure 2.
Figure 2

In this illustration, the lesions defined by the gold standard (circle by solid line) and general practitioner (GP; circle by dotted line) are areas B+C+D and D+E+F, respectively. The cropped image generated by GP (rectangle by dotted line) is presented as area C+D+E+F+G+H+I. The Jaccard index by definition is the area overlap percentage defined between by each GP and the gold standard, which is D/(B+C+D+E+F). The lesion inclusion rate is the percentage of ground-truth area included in the cropped image area generated by GP, that is, (C+D)/(B+C+D) in this illustration. With a high lesion inclusion rate, the main differences between each rectangular lesion derived from the marked borders by each GP and the gold standard may primarily involve background peripheral normal skin (F+G and A), which could be assumed to have similar characteristics.

In this illustration, the lesions defined by the gold standard (circle by solid line) and general practitioner (GP; circle by dotted line) are areas B+C+D and D+E+F, respectively. The cropped image generated by GP (rectangle by dotted line) is presented as area C+D+E+F+G+H+I. The Jaccard index by definition is the area overlap percentage defined between by each GP and the gold standard, which is D/(B+C+D+E+F). The lesion inclusion rate is the percentage of ground-truth area included in the cropped image area generated by GP, that is, (C+D)/(B+C+D) in this illustration. With a high lesion inclusion rate, the main differences between each rectangular lesion derived from the marked borders by each GP and the gold standard may primarily involve background peripheral normal skin (F+G and A), which could be assumed to have similar characteristics.

Statistics

The performance of CADx was evaluated based on receiver operating characteristic (ROC) curve analysis, with the pathological results considered as the gold standard of malignancy diagnosis. The area under the ROC curve (Az) was estimated after each physician using CADx. The Az values of two ROC curves were compared using DeLong's test.24 The masks generated from skin lesion segmentation by all seven physicians and JSEG were compared based on the Jaccard coefficient.23 The Wilcoxon rank sum test was used to compare the calculated Jaccard indices between the GPs and JSEG. Separate ICCs were computed to assess the contribution of different lesion borders determined by different readers to the 16 feature scores and the probability score. The ICC was calculated using a two-way random model with measures of absolute agreement.22 All statistical analyses were performed using R V.3.1.0.25

Results

Demographics

From January 2010 to December 2010, a total of 347 images of distinct regions of interest were obtained from 295 patients, including 124 males (42%) and 171 females (58%) with a mean age of 50.9±20.4 years. These images included 97 malignant lesions and 250 benign lesions. The demographic data associated with each histological diagnosis are summarised in table 1. There were seven invasive melanomas in this study. The number of melanomas in our database was small, consistent with the relatively low incidence in the Asian population compared with Caucasians.
Table 1

Demographic data for each histological diagnosis

PathologyNPer centSex (F/M)Mean age (year)
Benign25072.05156/9343.69
 Blue naevus123.469/342.67
 Compound naevus257.2018/632.52
 Dermatofibroma30.861/235.67
 Haemangioma133.757/653.00
 Intradermal naevus10931.4172/3737.84
 Junctional naevus216.0514/733.10
 Seborrheic keratosis6719.3135/3259.42
Malignant9727.9543/5469.75
 Basal cell carcinoma4713.5423/2471.17
 Bowen’s disease174.908/968.29
 Kaposi’s sarcoma41.150/448.25
 Keratoacanthoma30.861/256.00
 Melanoma72.025/263.86
 Squamous cell carcinoma195.486/1376.42
All lesions347100.00199/14750.97
Demographic data for each histological diagnosis

CADx performance

To compare the performance of CADx used by the dermatologists and GPs, the Az values based on individual manual segmentation by three dermatologists and four GPs were calculated and compared (figure 3). The border determination performance based on the gold standard was also assessed as previously described. The Az values of the CADx system for classification of benign or malignant skin lesions were as follows: 0.893 (95% CI 0.856 to 0.930) when segmented by the gold standard, 0.886 (95% CI 0.863 to 0.908) when segmented by the dermatologists, 0.883 (95% CI 0.864 to 0.903) when segmented by the GPs and 0.856 (95% CI 0.812 to 0.899) when segmented by JSEG. The Az values of the dermatologists were slightly better than those of the GPs, but there were no significant differences between the performances generated by the gold standard and the GPs (p=0.65). By adjusting the SKINCAD software operating point based on an optimal cut-off value by seven physicians at 0.3183, the sensitivity and specificity were 80.07% and 81.47%, respectively, for the dermatologists, which were comparable to the values of 79.90% and 80.12%, respectively, for the GPs. All physicians were able to complete the manual segmentation of each lesion in less than 1 min and the probability score could be generated by the CADx within 30 s. The average JSEG computation time is 8.9±8.1 s for each lesion. Owing to the complexity of the clinical images, an extra 10–15 s was required prior to the JSEG segmentation process, to manually preselect an area of interest.
Figure 3

The performance of the discrimination of skin malignancy by four ROC curves generated from segmentation results produced by the gold standard, dermatologists, GP and JSEG. Please note that the 34 failure cases of autosegmentation (9.8%) by JSEG are not included in the ROC analysis (AUC, area under curve; GP, general practitioner; ROC, receiver operating characteristic).

The performance of the discrimination of skin malignancy by four ROC curves generated from segmentation results produced by the gold standard, dermatologists, GP and JSEG. Please note that the 34 failure cases of autosegmentation (9.8%) by JSEG are not included in the ROC analysis (AUC, area under curve; GP, general practitioner; ROC, receiver operating characteristic).

Overlapping results

After omitting 34 (9.8%, 34/347) cases in which JSEG failed at border detection, 313 (90.2%) gold standard images were used to compare the JSEG and GP results (table 2). As the gold standard for overlap evaluation was derived from the dermatologists’ original markings, results of the dermatologists were not included for analysis.
Table 2

The Jaccard index and lesion inclusion rate of GP and JSEG, compared with the gold standard

PathologyJaccard index
Lesion inclusion rate
GPJSEGp ValueGPJSEGp Value
Benign0.68±0.150.60±0.260.000.96±0.100.86±0.300.00
 Blue naevus0.64±0.140.65±0.100.880.94±0.110.95±0.050.60
 Compound naevus0.70±0.140.62±0.230.100.96±0.090.92±0.210.12
 Dermatofibroma0.68±0.130.81±0.100.200.96±0.051.00±0.000.15
 Haemangioma0.65±0.200.63±0.240.820.99±0.030.96±0.100.32
 Junctional naevus0.61±0.160.58±0.270.020.91±0.150.83±0.340.00
 Intradermal naevus0.68±0.140.51±0.290.370.95±0.100.74±0.390.04
 Seborrheic keratosis0.72±0.130.62±0.280.070.97±0.110.87±0.280.00
Malignant0.73±0.150.60±0.280.000.98±0.080.84±0.320.00
 Basal cell carcinoma0.72±0.140.65±0.250.210.97±0.080.88±0.280.06
 Bowen's disease0.79±0.100.56±0.360.031.00±0.010.77±0.390.00
 Kaposi's sarcoma0.75±0.110.76±0.100.890.98±0.071.00±0.000.08
 Keratoacanthoma0.66±0.170.51±0.450.950.94±0.140.67±0.580.82
 Melanoma0.75±0.120.57±0.290.100.98±0.030.88±0.210.31
 Squamous cell carcinoma0.68±0.190.50±0.300.010.98±0.120.78±0.370.00
All lesions0.70±0.150.60±0.270.000.96±0.100.85±0.310.00

Note that 34 cases in which JSEG failed for border detection were not included in the analysis.

GP, general practitioner.

The Jaccard index and lesion inclusion rate of GP and JSEG, compared with the gold standard Note that 34 cases in which JSEG failed for border detection were not included in the analysis. GP, general practitioner. The overall Jaccard index of the lesions segmented by the GPs and JSEG compared with the gold standard was 0.70±0.15 and 0.60±0.27, respectively. The GPs performed better than JSEG with respect to the Jaccard index in the benign and malignant categories (p<0.01). When using cropped images for lesion inclusion rate (the percentage of the ground-truth lesion included in the cropped image area derived from user-defined borders), the overlap result using borders derived from GPs reached a higher inclusion ratio (0.96±0.10) compared with that derived from JSEG (0.85±0.31) for benign and malignant lesions (p<0.01).

Interobserver variability

The agreement associated with 16 individual features using different segmented areas generated by the seven physicians is summarised in table 3. These features are listed in order according to the recursive feature elimination ranking in our previous study.4 They are the foundation of CADx computation, and high agreement indicates CADx scoring consistency. The agreement in compactness and radial variance between the seven physicians was 0.57 and 0.63, respectively (fair-to-good level). The agreement regarding texture features, including grey level run length matrix and coarseness, was excellent (0.84–0.94). Colour features, for example, PC3, showed excellent agreement at a level of 0.96–0.98, and the conventional colour features also reached an excellent agreement level of 0.79–0.97 among all the physicians. The overall probability score derived from the seven physicians reached excellent agreement (0.91). We also investigated the agreement in feature scores between the individual GPs and the gold standard derived from dermatologists. All GPs reached excellent agreement in the final probability score and all 14 features (0.77–0.99), with the exception of compactness and radial variance (0.52–0.75), whereas there was only fair agreement (0.48) between JSEG and the gold standard in the final probability score, and poor agreement in the compactness and radial variance features at 0.13 and 0.29, respectively.
Table 3

The agreement scores of each feature generated from segmentation results by all users, JSEG and the gold standard, assessed by ICC

FeaturesAll 7 physiciansGP 1 vs goldGP 2 vs goldGP 3 vs goldGP 4 vs goldJSEG vs gold
PC3*0.960.970.970.990.980.95
Variance blue channel*0.880.890.900.950.920.84
Variance blue channel†0.970.970.970.990.980.98
Compactness*0.570.520.570.700.640.13
Radial variance*0.630.670.600.750.580.29
Green–blue correlation*0.770.870.770.800.870.65
Green–grey correlation*0.800.880.820.880.870.66
PC3†0.980.980.970.990.990.95
Entropy red channel†0.900.940.890.950.900.93
Entropy red channel*0.920.960.900.970.940.86
Entropy blue channel†0.930.930.940.970.960.94
Entropy blue channel*0.880.930.880.940.900.82
GLRLM_HGRE_4Level†0.840.910.770.930.870.86
GLRLM_SRLGE_4Level†0.950.960.950.980.970.93
GLRLM_SRLGE_2Level†0.940.940.940.970.950.94
Tamura's coarseness features*0.940.930.930.970.950.91
Probability score0.910.910.900.940.930.48

Features are listed in order according to RFE ranking between 91 features in our previous study. All of the p values of each ICCs in this table are ≤0.01. The failure cases (34/347) of autosegmentation by JSEG are not included in the analysis.

*Derived from the lesion area only.

†Derived from the whole cropped image.

GLRLM, grey level run length matrix; GP, general practitioner; HGRE, high grey level run emphasis; ICC, intraclass correlation coefficient; PC3, the variance along the coordinates of the third principal components; RFE, recursive feature elimination; SRLGE, short run low grey level emphasis.

The agreement scores of each feature generated from segmentation results by all users, JSEG and the gold standard, assessed by ICC Features are listed in order according to RFE ranking between 91 features in our previous study. All of the p values of each ICCs in this table are ≤0.01. The failure cases (34/347) of autosegmentation by JSEG are not included in the analysis. *Derived from the lesion area only. †Derived from the whole cropped image. GLRLM, grey level run length matrix; GP, general practitioner; HGRE, high grey level run emphasis; ICC, intraclass correlation coefficient; PC3, the variance along the coordinates of the third principal components; RFE, recursive feature elimination; SRLGE, short run low grey level emphasis.

Discussion

Precise segmentation was considered an essential step when using the CADx system for skin cancer diagnosis.4 Without the use of microscopic facilities, there were no definite criteria for defining the true borders between the lesion and non-lesion areas under gross examination, even by expert dermatologists.26 Therefore, subjectivity is always a concern, and automated segmentation may not always be applicable to these clinical situations. There remained no unified results available for comparing all the tested segmentation algorithms due to differences in the ground-truth definitions and evaluation metrics.27–29 When utilising conventional digital photographs for the analysis of melanocytic and non-melanocytic lesions, the images usually consist of more colours than the dermoscopic images of melanocytic lesions. Regarding clinical application, subjective manual segmentation in SKINCAD is especially useful for images with complicated components, as human eyes are good at pattern recognition for diagnosis, whereas automated segmentation methods sometimes fail. The failure rate of JSEG was 9.8% (34/347) in our study, similar to the situation of a previous report based on dermoscopic images in real clinical settings in which cases may be rejected for analysis by CADx due to unsuccessful autosegmentation.14 In addition, our software provides an easy-to-use interface for manual segmentation. Users are able to complete the process in less than 2 min for each lesion, including the manual segmentation process and the probability score computation. The strength of this study lies in the fact that the groups of GPs and dermatologists have both manually segmented a wide spectrum of skin lesions under generous inclusion criteria, which represents skin lesions encountered in daily practice, with each lesion given a definite histopathological diagnosis. SKINCAD performed as well as face-to-face clinical diagnosis by staff dermatologists at our institution in our previous report.4 In this study, with a new data set unknown to all raters and SKINCAD, SKINCAD achieved good Az performance for classification of benign or malignant skin lesions using borders determined by either GPs, dermatologists or the gold standard developed from dermatologists’ original markings. Instead of testing segmentation methods to merely approximate the ground-truth tumour area determined by dermatologists, which may be of concern as standard, we successfully validated the system's reliability with melanocytic and non-melanocytic skin lesion classification based on a good-sized data set with real clinical settings by assessing the robustness with respect to each feature and final probability score agreement levels derived from borders segmented by individual users of different backgrounds. We introduced the agreement in 16 features as a performance metric. SKINCAD performed well with colour features showing excellent agreement (0.77–0.98) between the two groups of physicians. The segmentation differences between the users resulted in discrepancies mainly with respect to the border-derived features of compactness and radial variance; only fair-to-good agreement was reached for these features. Given the known individual subjectivity of each physician, and the results of agreement with the dermatologist-derived gold standard regarding all features and the final probability score, GPs reached a stable performance that was better than that achieved by JSEG. Generally, human users performed better than JSEG in all overlap indices. As 7 of 14 colour and texture features were derived from the cropped rectangular images generated from each crooked border determined by the users, the images for analysis consisted of the main tumour lesion area and peripheral background skin during the preprocessing of the peripheral extension. Therefore, the lesion inclusion rate, which was 0.96±0.1 in this study, also contributed to the stability of the analysis results. This result indicates that a very high proportion of the main lesion on average was included in the cropped rectangular images used for analysis in spite of the discrepancies in the borders drawn by each GP. With a high lesion inclusion rate, the main differences between each rectangular lesion derived from the marked borders by each GP and the gold standard may primarily involve background peripheral normal skin, which could be assumed to have similar characteristics (figure 2). This result may also explain that despite an average Jaccard index of 0.70±0.15 between the GPs and the gold standard, the evaluations by all of the physicians still achieved excellent agreement with respect to most features and the final probability scores. A consistent performance in colour features was maintained with subjective manual segmented border variation. This implies that when other useful features, besides shape features, were selected in the SKINCAD, the segmentation variation related to borders may have impacted less on the classification accuracy than a system that uses border-sensitive features only. There were limitations in this study. All images were obtained from patients visiting a single centre in southern Taiwan using a single image-capture system with consistent quality control for each photograph. The analysis was retrospective and restricted to biopsied lesions, as we used the histopathological reports as the gold standard. We were unable to evaluate the performance regarding lesions for which clinicians or patients decided not to perform the biopsy, as they were not included in the data set. The performance of this system was not compared with that in other CADx studies due to different clinical settings. Further large-scale prospective study may be required in the future for broader application. In conclusion, our study established a possible model for the diagnosis of skin lesions using conventional digital photography with manual segmentation. A system with multiple features, including border-sensitive and non-border-sensitive features, may compensate for the impact of the subjectivity of manual segmentation. This may be an appropriate solution especially when automatic segmentation is not feasible or applicable. Through the in-depth evaluation of overlap index and feature agreement levels, our study indicates the possibility of direct onsite computation application for physicians other than dermatology specialists, when assessing skin lesions. By combining effective feature extractions by modern computation technologies, and manual segmentations of the lesion area and peripheral skin, SKINCAD may play a role as a consistent second opinion for dermatologists and for GPs. Research on the benefits of SKINCAD with respect to clinical decision-making improvement should be performed in the future.
  19 in total

1.  Validation of segmentation techniques for digital dermoscopy.

Authors:  Guillod Joel; Philippe Schmid-Saugeon; David Guggisberg; Jean Philippe Cerottini; Ralph Braun; Joakim Krischer; Jean-Hilaire Saurat; Kunt Murat
Journal:  Skin Res Technol       Date:  2002-11       Impact factor: 2.365

2.  Towards a computer-aided diagnosis system for pigmented skin lesions.

Authors:  Philippe Schmid-Saugeona; Joël Guillodb; Jean-Philippe Thirana
Journal:  Comput Med Imaging Graph       Date:  2003       Impact factor: 4.790

3.  Border detection in dermoscopy images using hybrid thresholding on optimized color channels.

Authors:  Rahil Garnavi; Mohammad Aldeen; M Emre Celebi; George Varigos; Sue Finch
Journal:  Comput Med Imaging Graph       Date:  2010-09-15       Impact factor: 4.790

4.  Weighted performance index for objective evaluation of border detection methods in dermoscopy images.

Authors:  Rahil Garnavi; Mohammad Aldeen; M E Celebi
Journal:  Skin Res Technol       Date:  2011-02       Impact factor: 2.365

Review 5.  Computerized analysis of pigmented skin lesions: a review.

Authors:  Konstantin Korotkov; Rafael Garcia
Journal:  Artif Intell Med       Date:  2012-10-11       Impact factor: 5.326

6.  A perceptually oriented method for contrast enhancement and segmentation of dermoscopy images.

Authors:  Qaisar Abbas; Irene Fondón Garcia; M Emre Celebi; Waqar Ahmad; Qaisar Mushtaq
Journal:  Skin Res Technol       Date:  2012-08-13       Impact factor: 2.365

7.  Can automated dermoscopy image analysis instruments provide added benefit for the dermatologist? A study comparing the results of three systems.

Authors:  A Perrinaud; O Gaide; L E French; J-H Saurat; A A Marghoob; R P Braun
Journal:  Br J Dermatol       Date:  2007-09-13       Impact factor: 9.302

Review 8.  Lesion border detection in dermoscopy images.

Authors:  M Emre Celebi; Hitoshi Iyatomi; Gerald Schaefer; William V Stoecker
Journal:  Comput Med Imaging Graph       Date:  2009-01-03       Impact factor: 4.790

9.  Analysis of skin cancer treatment and costs in the United States Medicare population, 1996-2008.

Authors:  Howard W Rogers; Brett M Coldiron
Journal:  Dermatol Surg       Date:  2012-11-30       Impact factor: 3.398

10.  An improved border detection in dermoscopy images for density based clustering.

Authors:  Sait Suer; Sinan Kockara; Mutlu Mete
Journal:  BMC Bioinformatics       Date:  2011-10-18       Impact factor: 3.169

View more
  1 in total

Review 1.  The Possibility of Deep Learning-Based, Computer-Aided Skin Tumor Classifiers.

Authors:  Yasuhiro Fujisawa; Sae Inoue; Yoshiyuki Nakamura
Journal:  Front Med (Lausanne)       Date:  2019-08-27
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.