| Literature DB >> 31598159 |
Qing Guan1,2, Yunjun Wang1,2, Bo Ping2,3, Duanshu Li1,2, Jiajun Du4, Yu Qin4, Hongtao Lu4, Xiaochun Wan2,3, Jun Xiang1,2.
Abstract
Objective: In this study, we exploited a VGG-16 deep convolutional neural network (DCNN) model to differentiate papillary thyroid carcinoma (PTC) from benign thyroid nodules using cytological images.Entities:
Keywords: Deep convolutional neural network; cytological images; fine-needle aspiration; liquid-based cytology; papillary thyroid carcinoma
Year: 2019 PMID: 31598159 PMCID: PMC6775529 DOI: 10.7150/jca.28769
Source DB: PubMed Journal: J Cancer ISSN: 1837-9664 Impact factor: 4.207
Figure 1Fragmented cytological images. (A, B, C): PTC; (D, E, F): benign nodules.
Number of fragmented images in the dataset.
| Cytological type | Training data | Test data | Total |
|---|---|---|---|
| PTC | 407 | 69 | 476 |
| Benign nodule | 352 | 59 | 411 |
| All type | 759 | 128 | 887 |
Figure 3Quantification of tumor cells. (A) Original images, (B) Grayscale images, (C) contours of tumor cells.
Architecture of VGG-16 network
| Layer | Patch size | Input size |
|---|---|---|
| conv×2 | 3×3/1 | 3×224×224 |
| pool | 2×2 | 64×224×224 |
| conv×2 | 3×3/1 | 64×112×112 |
| pool | 2×2 | 128×112×112 |
| conv×3 | 3×3/1 | 128×56×56 |
| pool | 2×2 | 256×56×56 |
| conv×3 | 3×3/1 | 256×28×28 |
| pool | 2×2 | 512×28×28 |
| conv×3 | 3×3/1 | 512×14×14 |
| pool | 2×2 | 512×14×14 |
| fc | 25088×4096 | 25088 |
| fc | 4096×4096 | 4096 |
| fc | 4096×2 | 4096 |
| softmax | classifier | 2 |
Conv stands for convolutional layer, pool stands for pooling layer and fc stands for fully connected layer. Patch size is the kernel size of convolutional layer, pooling layer or fully connected layer. Input size is feature map input size of the layer.
Architecture of Inception-v3 network
| Layer | Patch size | Input size |
|---|---|---|
| conv | 3×3/2 | 229×229×3 |
| conv | 3×3/1 | 149×149×32 |
| conv padded | 3×3/1 | 147×147×32 |
| pool | 3×3/2 | 147×147×64 |
| conv | 3×3/1 | 73×73×64 |
| conv | 3×3/2 | 71×71×80 |
| conv | 3×3/1 | 35×35×192 |
| Inception A×3 | ------ | 35×35×288 |
| Inception B×5 | ------ | 17×17×768 |
| Inception C×2 | ------ | 8×8×1280 |
| pool | 8×8 | 8×8×2048 |
| linear | logits | 2048 |
| softmax | classifier | 2 |
Conv stands for convolutional layer, pool stands for pooling layer and fc stands for fully connected layer. Patch size is the kernel size of convolutional layer, pooling layer or fully connected layer. Input size is feature map input size of the layer.
Figure 2The Inception modules of Inception-v3. The Inception modules of Inception-v3 including Inception module A, B and C from left to right. Each Inception module is composed of several convolutional layers and pooling layers. Pool stands for pooling layer and n×m stands for n×m convolutional layer.
Diagnostic efficiency of VGG-16 and Inception-v3 on test data (fragmented images)
| Model | VGG-16 | Inception-v3 |
|---|---|---|
| Accuracy | 97.66% | 92.75% |
| Sensitivity | 100% | 98.55% |
| Specificity | 94.91% | 86.44% |
| Positive predictive value | 95.83% | 89.47% |
| Negative predictive value | 100% | 98.08% |
Quantification of tumor cells in fragmented images of malignant and benign thyroid tumors.
| Malignant | Benign | p value | |
|---|---|---|---|
| Contour | 61.01±17.10 | 47.00±24.08 | < 0.001 |
| Perimeter | 134.99±21.42 | 62.40±29.15 | < 0.001 |
| Area | 1770.89±627.22 | 1157.27±722.23 | 0.013 |
| Mean of pixel intensity | 165.84±26.33 | 132.94±28.73 | < 0.001 |
Figure 4Misdiagnosed fragmented images. (B) and (C) were cropped from same image. A: contour 17, perimeter 149.67, area 1685.91, mean pixel intensity 163.47; B: Contour 19, perimeter 107.17, area 1469.83, mean pixel intensity 165.35; C: Contour 21, perimeter 127.00, area 1839.65, mean pixel intensity 182.79.