| Literature DB >> 34455613 |
Qiang Lin1,2,3, Chuangui Cao1,2, Tongtong Li1,2, Yongchun Cao1,2,3, Zhengxing Man1,2,3, Haijun Wang4.
Abstract
PURPOSE: A self-defined convolutional neural network is developed to automatically classify whole-body scintigraphic images of concern (i.e., the normal, metastasis, arthritis, and thyroid carcinoma), automatically detecting diseases with whole-body bone scintigraphy.Entities:
Keywords: attention mechanism; bone scintigraphy; convolutional neural network; medical image analysis; multiclass classification
Mesh:
Year: 2021 PMID: 34455613 PMCID: PMC9135133 DOI: 10.1002/mp.15196
Source DB: PubMed Journal: Med Phys ISSN: 0094-2405 Impact factor: 4.506
Number of diseases in patients involved in the collected SPECT scintigraphic images
| Normal | Bone metastasis | Arthritis | Thyroid carcinoma | |
|---|---|---|---|---|
| Patient | 179 | 117 | 143 | 161 |
| Proportion | 30% | 20% | 24% | 26% |
An overview of our SPECT scintigraphic images
| Normal | Bone metastasis | Arthritis | Thyroid carcinoma | |
|---|---|---|---|---|
| Number of images | 334 | 174 | 252 | 318 |
| Proportion | 31% | 17% | 23% | 29% |
FIGURE 1Illustration of mirroring, translating, and rotating whole‐body SPECT scintigraphic image. (a) Original posterior image; (b) mirrored image; (c) translated image; and (d) rotated image by 3o to the right direction
An overview of the augmented dataset of SPECT scintigraphic images
| Normal | Bone metastasis | Arthritis | Thyroid carcinoma | |
|---|---|---|---|---|
| Number of images | 1660 | 1582 | 1500 | 1788 |
| Proportion | 26% | 24% | 23% | 27% |
FIGURE 2Illustration of labelling a 2D whole‐body SPECT scintigraphic image using the LabelMe‐based annotation system
Structure and parameters of the self‐defined deep classification network Dscint
| Layer | Configuration |
|---|---|
| Convolution | 11 × 11, 16, S = 4, P = 2 |
| Pooling | MaxPool(3), S = 2 |
| Attention module | |
| Convolution | 5 × 5, 16, S = 1, P = 2 |
| BatchNorm | |
| Pooling | MaxPool(3), S = 2 |
| Convolution | 3 × 3, 24, S = 1, P = 1 |
| Convolution | 3 × 3, 24, S = 1, P = 1 |
| Convolution | 3 × 3, 24, S = 1, P = 1 |
| BatchNorm | |
| Pooling | MaxPool(3), S = 1 |
| Fully connected | 1024 |
| Fully connected | 1024 |
| Softmax | 4 |
Abbreviations: MaxPool, max pooling; P, padding; S, stride.
FIGURE 3Hybrid attention module with the channel and spatial attention in the self‐defined Dscint network
Parameter settings of the self‐defined classification network of Dscint
| Parameter | Value |
|---|---|
| Learning rate | 10–3 |
| Weight decay | 10–4 |
| Batch size | 4 |
| Epoch | 300 |
| Iteration | 1200 |
FIGURE 4Illustration of training Dscint on the original (blue) and augmented (orange) datasets. (a) Accuracy curves and (b) loss curves
Scores of evaluation metrics obtained by Dscint on the test samples in both original and augmented dataset
| Accuracy | Precision | Recall | Specificity |
| |
|---|---|---|---|---|---|
| Original data | 0.8519 | 0.8599 | 0.8257 | 0.9489 | 0.8362 |
| Augmented data | 0.9801 | 0.9795 | 0.9791 | 0.9933 | 0.9792 |
FIGURE 5Quantitative performance obtained by Dscint on test samples in the augmented dataset with average scores of evaluation metrics for different classes of concerns
FIGURE 6Examples of misclassified whole‐body SPECT scintigraphic images with N = Normal; M = Metastasis; A = Arthritis; and T = Thyroid carcinoma. (a) Correctly classified images and (b) wrongly classified images
An overview of the classical CNNs‐based models used for comparative analysis
| Weight layer | Filter | Activation | Optimizer | |
|---|---|---|---|---|
| AlexNet | 8 | 11 × 11, 5 × 5, 3 × 3 | ReLU | Adam |
| ResNet | 18 | 3 × 3 | ReLU | Adam |
| VGG‐16 | 16 | 3 × 3 | ReLU | Adam |
| Inception‐v4 | 14 Inception | 3 × 3, 1 × 1, 1 × 7, 7 × 1, 1 × 3, 3 × 1 | ReLU | Adam |
| DenseNet | 121 | 1 × 1, 3 × 3 | ReLU | Adam |
Evaluation metrics obtained by six models on test samples
| AlexNet | ResNet | VGG‐16 | Inception‐v4 | DenseNet | Dscint | |
|---|---|---|---|---|---|---|
| Accuracy | 0.9652 | 0.9314 | 0.9550 | 0.8384 | 0.9371 | 0.9801 |
| Precision | 0.9641 | 0.9316 | 0.9538 | 0.8367 | 0.9391 | 0.9795 |
| Recall | 0.9643 | 0.9293 | 0.9556 | 0.8348 | 0.9332 | 0.9791 |
| Specificity | 0.9884 | 0.9767 | 0.9850 | 0.9420 | 0.9781 | 0.9933 |
| F‐1 score | 0.9715 | 0.9303 | 0.9541 | 0.8345 | 0.9309 | 0.9792 |
FIGURE 7A comparison of evaluation metrics obtained by CNNs‐based classification models on test samples in the augmented dataset. (a) Specificity and (b) F‐1 score
FIGURE 8ROC curves obtained by six models on test samples of the augmented dataset in Table 3
AUC values obtained by six models on test samples of the augmented dataset in Table 3
| AlexNet | ResNet | VGG‐16 | Inception‐v4 | DenseNet | Dscint | |
|---|---|---|---|---|---|---|
| AUC | 0.9973 | 0.9877 | 0.9921 | 0.9532 | 0.9838 | 0.9985 |
FIGURE 9Confusion matrices obtained by six models on test samples in the augmented dataset