| Literature DB >> 28695342 |
Jianning Chi1, Ekta Walia2, Paul Babyn3, Jimmy Wang3, Gary Groot4, Mark Eramian2.
Abstract
With many thyroid nodules being incidentally detected, it is important to identify as many malignant nodules as possible while excluding those that are highly likely to be benign from fine needle aspiration (FNA) biopsies or surgeries. This paper presents a computer-aided diagnosis (CAD) system for classifying thyroid nodules in ultrasound images. We use deep learning approach to extract features from thyroid ultrasound images. Ultrasound images are pre-processed to calibrate their scale and remove the artifacts. A pre-trained GoogLeNet model is then fine-tuned using the pre-processed image samples which leads to superior feature extraction. The extracted features of the thyroid ultrasound images are sent to a Cost-sensitive Random Forest classifier to classify the images into "malignant" and "benign" cases. The experimental results show the proposed fine-tuned GoogLeNet model achieves excellent classification performance, attaining 98.29% classification accuracy, 99.10% sensitivity and 93.90% specificity for the images in an open access database (Pedraza et al. 16), while 96.34% classification accuracy, 86% sensitivity and 99% specificity for the images in our local health region database.Entities:
Keywords: Computer vision; Convolutional neural network; Deep learning; Fine-tuning; Machine learning; Thyroid nodules; Ultrasonography
Mesh:
Year: 2017 PMID: 28695342 PMCID: PMC5537102 DOI: 10.1007/s10278-017-9997-y
Source DB: PubMed Journal: J Digit Imaging ISSN: 0897-1889 Impact factor: 4.056
Fig. 1Example of artifacts made by radiologist on the thyroid ultrasound image. a Ultrasound image with artifacts covering the textures. b Details of how the artifact covers the ultrasound textures in the image, bounded by the red rectangle in a
Fig. 2The process of the thyroid images classification based on fine-tuned GoogLeNet network
Fig. 3Detection of pixels deemed to be ticks by intensity thresholding. a Input thyroid ultrasound image. b Thirty percent right region of the input image where the thresholding was applied. c Binary image showing the pixels deemed likely to be ticks, and red line in d represents the column containing the tick bar
Fig. 4Example of removing artifacts from the thyroid image. a Original thyroid image with artifacts on the textures. b Thyroid image after removing the artifacts by the thresholding method. c Thyroid image after restoring the artifact-removed regions with textures similar to neighbourhood
The distribution of positive and negative cases in training, validating and testing groups of database 1
| Total | Positive | Negative | ||||
|---|---|---|---|---|---|---|
| Cases | Samples | Cases | Samples | Cases | Samples | |
| Training | 306 | 2754 | 256 | 2304 | 50 | 450 |
| Validating | 61 | 549 | 51 | 459 | 10 | 90 |
| Testing | 61 | 549 | 50 | 450 | 11 | 99 |
The distribution of positive and negative cases in training, validating and testing groups of database 2
| Total | Positive | Negative | ||||
|---|---|---|---|---|---|---|
| Cases | Samples | Cases | Samples | Cases | Samples | |
| Training | 132 | 1188 | 21 | 189 | 111 | 999 |
| Validating | 16 | 144 | 4 | 36 | 12 | 108 |
| Testing | 16 | 144 | 4 | 36 | 12 | 108 |
Fig. 5The process of fine-tuning GoogLeNet. The new sample images from the target domain are sent to the pre-trained GoogLeNet, changed the three classification layers and run the GoogLeNet, and then the parameters of the network are fine-tuned automatically
The results of classifying testing samples from database 1 by the proposed model (GoogLeNet fine-tuned by samples from both database 1 and database 2)
| Accuracy | Sensitivity | Specificity | AUC | |
|---|---|---|---|---|
| Proposed model | 99.13% | 99.70% | 95.80% | 0.9970 |
Fig. 6The ROC curves of classifying image samples from database 1
The results of classifying testing samples from database 2 by the proposed model (GoogLeNet fine-tuned by samples from both database 1 and database 2)
| Accuracy | Sensitivity | Specificity | AUC | |
|---|---|---|---|---|
| Proposed model | 96.34% | 82.80% | 99.30% | 0.9920 |
Fig. 7The ROC curves of classifying image samples from database 2
Comparison of different CAD system in classifying benign and malignant thyroid nodule images
| Method | Feature | Machine learning | Testing samples | Accuracy |
|---|---|---|---|---|
| Lim et al. [ | Size, margin, echogenicity, cystic change | Artificial neural network | 190 thyroid lesions | 93.78% |
| Savelonas et al. [ | Shape features | Nearest neighbour (k-NN) | 173 longitudinal in vivo images | 93.00% |
| Iakovidis et al. [ | Fuzzy intensity histogram | SVM | 250 thyroid ultrasound images | 97.50% |
| Legakis et al. [ | Texture features, shape features | SVM | 142 longitudinal in vivo images | 93.20% |
| Luo et al. [ | Strain rate waveform’s power spectrum | Linear discriminant analysis | 98 nodule image sequences | 87.50% |
| Acharya et al. [ | Higher-order spectra features | Fuzzy classifier | 80 sample 3D images | 99.10% |
| Proposed model | Deep learning features | Cost-sensitive Random Forest classifier | 693 sample images | 99.13% |