| Literature DB >> 34188673 |
Minghui Guo1, Kangjian Wang2, Shunlan Liu3, Yongzhao Du1,4,5, Peizhong Liu1,4,5, Qichen Su3,5, Guorong Lv3,5.
Abstract
Ultrasound is one of the critical methods for diagnosis and treatment in thyroid examination. In clinical application, many reasons, such as large outpatient traffic, time-consuming training of sonographers, and uneven professional level of physicians, often cause irregularities during the ultrasonic examination, leading to misdiagnosis or missed diagnosis. In order to standardize the thyroid ultrasound examination process, this paper proposes using a deep learning method based on residual network to recognize the Thyroid Ultrasound Standard Plane (TUSP). At first, referring to multiple relevant guidelines, eight TUSP were determined with the advice of clinical ultrasound experts. A total of 5,500 TUSP images of 8 categories were collected with the approval and review of the Ethics Committee and the patient's informed consent. Then, after desensitizing and filling the images, the 18-layer residual network model (ResNet-18) was trained for TUSP image recognition, and five-fold cross-validation was performed. Finally, through indicators like accuracy rate, we compared the recognition effect of other mainstream deep convolutional neural network models. Experimental results showed that ResNet-18 has the best recognition effect on TUSP images with an average accuracy rate of 91.07%. The average macro precision, average macro recall, and average macro F1-score are 91.39%, 91.34%, and 91.30%, respectively. It proves that the deep learning method based on residual network can effectively recognize TUSP images, which is expected to standardize clinical thyroid ultrasound examination and reduce misdiagnosis and missed diagnosis.Entities:
Year: 2021 PMID: 34188673 PMCID: PMC8192196 DOI: 10.1155/2021/5598001
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1Thyroid Ultrasound Standard Plane images. (TI represents thyroid isthmus, LT and RT represent the left and right lobe of the thyroid, respectively, T represents the trachea, and ES represents esophagus). (a) TPTI, (b) LPTI, (c) UTPLT, (d) DTPLT, (e) UTPRT, (f) DTPRT, (g) LPLT, and (h) LPRT.
Figure 2The structure of the residual block.
Figure 3The specific structure of the ResNet-18 model.
Architecture of ResNet-18.
| Block | Layers | Output size |
|---|---|---|
| Input | Input layer | 500 × 500 × 1 |
| Conv 1 | 7 | 250 × 250 × 64 |
| Pooling | 2 | 125 × 125 × 64 |
| Residual block 1 | 3 | 125 × 125 × 64 |
| 3 | ||
| Residual block 2 | 3 | 125 × 125 × 64 |
| 3 | ||
| Residual block 3 | 3 | 63 × 63 × 128 |
| 3 | ||
| Residual block 4 | 3 | 63 × 63 × 128 |
| 3 | ||
| Residual block 5 | 3 | 32 × 32 × 256 |
| 3 | ||
| Residual block 6 | 3 | 32 × 32 × 256 |
| 3 | ||
| Residual block 7 | 3 | 16 × 16 × 512 |
| 3 | ||
| Residual block 8 | 3 | 16 × 16 × 512 |
| 3 × 3 conv | ||
| Avg pooling | 16 × 16 avg pooling | 1 × 1 × 512 |
| FC layer | FC softmax | 1 × 1 × 8 |
Distribution of 8 categories of TUSP images.
| Types of TUSP | Number | Types of TUSP | Number |
|---|---|---|---|
| TPTI | 635 | UTPRT | 586 |
| LPTI | 1002 | DTPRT | 489 |
| UTPLT | 583 | LPLT | 733 |
| DTPLT | 552 | LPRT | 920 |
| Sum | 5500 | ||
Figure 4Padding and resizing of TUSP images.
The precision, recall, and F1 score of various categories in the test set.
| Types of TUSP | Precision | Recall |
|
|---|---|---|---|
| TPTI | 0.9815 | 0.9953 | 0.9883 |
| LPTI | 0.9980 | 0.9970 | 0.9975 |
| UTPLT | 0.9410 | 0.9399 | 0.9401 |
| DTPLT | 0.9328 | 0.9241 | 0.9277 |
| UTPRT | 0.9512 | 0.9146 | 0.9322 |
| DTPRT | 0.9046 | 0.9408 | 0.9218 |
| LPLT | 0.7852 | 0.7680 | 0.7753 |
| LPRT | 0.8170 | 0.8272 | 0.8212 |
| Macro average |
|
|
|
| Accuracy |
| ||
The values in the table are the average of five-fold cross-validation.
Figure 5The confusion matrix of the experiment result (the values in the figure are the average of five-fold cross-validation).
Recognition effects of different CNN models on TUSP images.
| Models | TPTI | LPTI | UTPLT | DTPLT | UTPRT | DTPRT | LPLT | LPRT | Macro average | Accuracy | |
|---|---|---|---|---|---|---|---|---|---|---|---|
| ResNet-18 |
| 0.9815 |
|
| 0.9328 |
| 0.9046 |
|
|
|
|
|
|
| 0.9970 | 0.9399 | 0.9241 | 0.9146 |
|
| 0.8272 |
| ||
|
| 0.9883 |
| 0.9401 | 0.9277 |
| 0.9218 |
|
|
| ||
|
| |||||||||||
| ResNet-50 |
| 0.9845 | 0.9832 | 0.9026 | 0.8956 | 0.9179 | 0.8792 | 0.6884 | 0.7812 | 0.8791 | 0.8744 |
|
| 0.9843 | 0.9890 | 0.9004 | 0.8842 | 0.8909 | 0.9100 | 0.7329 | 0.7348 | 0.8783 | ||
|
| 0.9843 | 0.9861 | 0.9006 | 0.8890 | 0.9030 | 0.8929 | 0.7078 | 0.7558 | 0.8775 | ||
|
| |||||||||||
| ResNet-101 |
| 0.9922 | 0.9902 | 0.9117 | 0.9281 | 0.9040 | 0.8964 | 0.7492 | 0.7242 | 0.8870 | 0.8795 |
|
| 0.9906 | 0.9910 | 0.9296 | 0.9041 | 0.9061 | 0.8876 | 0.6127 | 0.8261 | 0.8810 | ||
|
| 0.9913 | 0.9906 | 0.9203 | 0.9156 | 0.9048 | 0.8909 | 0.6710 | 0.7703 | 0.8818 | ||
|
| |||||||||||
| VGG16 |
| 0.9937 | 0.9851 | 0.9328 | 0.8929 | 0.9039 | 0.8798 | 0.7190 | 0.7461 | 0.8817 | 0.8762 |
|
| 0.9874 | 0.9900 | 0.8971 | 0.9221 | 0.8911 | 0.8813 | 0.6768 | 0.7815 | 0.8784 | ||
|
| 0.9905 | 0.9875 | 0.9132 | 0.9057 | 0.8956 | 0.8784 | 0.6963 | 0.7626 | 0.8787 | ||
|
| |||||||||||
| ResNet-152 |
|
| 0.9813 | 0.9247 | 0.9028 | 0.8826 | 0.8538 | 0.7800 | 0.7024 | 0.8777 | 08634 |
|
| 0.9858 | 0.9910 | 0.9057 |
| 0.8768 | 0.8569 | 0.5289 |
| 0.8635 | ||
|
| 0.9897 | 0.9861 | 0.9143 | 0.9128 | 0.8773 | 0.8507 | 0.6042 | 0.7556 | 0.8613 | ||
|
| |||||||||||
| InceptionV3 |
| 0.9907 | 0.9911 | 0.9359 | 0.9341 | 0.9220 | 0.8960 | 0.7430 | 0.7926 | 0.9007 | 0.8962 |
|
| 0.9858 | 0.9920 | 0.9398 | 0.9313 | 0.9098 | 0.9140 | 0.7407 | 0.7870 | 0.9000 | ||
|
| 0.9881 | 0.9915 | 0.9374 |
| 0.9150 | 0.9038 | 0.7398 | 0.7882 | 0.8995 | ||
|
| |||||||||||
| MobileNet |
| 0.9892 | 0.9921 | 0.9341 | 0.9294 | 0.9162 |
| 0.7490 | 0.8028 | 0.9031 | 0.8986 |
|
| 0.9937 |
| 0.9347 | 0.9223 |
| 0.8936 | 0.7613 | 0.7880 | 0.9012 | ||
|
|
| 0.9940 | 0.9340 | 0.9254 | 0.9174 | 0.9018 | 0.7528 | 0.7932 | 0.9012 | ||
|
| |||||||||||
| Xception |
| 0.9844 | 0.9913 | 0.9298 |
| 0.9504 | 0.9054 | 0.7634 | 0.7812 | 0.9083 | 0.9013 |
|
| 0.9890 | 0.9900 |
| 0.9061 | 0.9148 | 0.9406 | 0.7015 | 0.8315 | 0.9047 | ||
|
| 0.9867 | 0.9906 |
| 0.9319 | 0.9318 |
| 0.7258 | 0.8030 | 0.9047 | ||
The values in the table are the average of five-fold cross-validation.
The computational cost of different CNN models.
| CNN models | Trainable parameters | Nontrainable parameters | Total parameters |
|---|---|---|---|
| ResNet18 | 11,177,352 | 7,808 | 11,185,160 |
| ResNet50 | 23,544,712 | 53,120 | 23,597,832 |
| ResNet101 | 42,562,952 | 105,344 | 42,668,296 |
| ResNet152 | 58,229,640 | 151,424 | 58,381,064 |
| VGG16 | 134,292,168 | 0 | 134,292,168 |
| InceptionV3 | 21,784,168 | 34,432 | 21,818,600 |
| MobileNet | 3,214,600 | 21,888 | 3,236,488 |
| Xception | 20,822,768 | 54,528 | 20,877,296 |