| Literature DB >> 35918498 |
Shintaro Sukegawa1,2, Futa Tanaka3, Keisuke Nakano4, Takeshi Hara3,5, Kazumasa Yoshii3, Katsusuke Yamashita6, Sawako Ono7, Kiyofumi Takabatake4, Hotaka Kawai4, Hitoshi Nagatsuka4, Yoshihiko Furuki8.
Abstract
The use of sharpness aware minimization (SAM) as an optimizer that achieves high performance for convolutional neural networks (CNNs) is attracting attention in various fields of deep learning. We used deep learning to perform classification diagnosis in oral exfoliative cytology and to analyze performance, using SAM as an optimization algorithm to improve classification accuracy. The whole image of the oral exfoliation cytology slide was cut into tiles and labeled by an oral pathologist. CNN was VGG16, and stochastic gradient descent (SGD) and SAM were used as optimizers. Each was analyzed with and without a learning rate scheduler in 300 epochs. The performance metrics used were accuracy, precision, recall, specificity, F1 score, AUC, and statistical and effect size. All optimizers performed better with the rate scheduler. In particular, the SAM effect size had high accuracy (11.2) and AUC (11.0). SAM had the best performance of all models with a learning rate scheduler. (AUC = 0.9328) SAM tended to suppress overfitting compared to SGD. In oral exfoliation cytology classification, CNNs using SAM rate scheduler showed the highest classification performance. These results suggest that SAM can play an important role in primary screening of the oral cytological diagnostic environment.Entities:
Mesh:
Year: 2022 PMID: 35918498 PMCID: PMC9346110 DOI: 10.1038/s41598-022-17602-4
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.996
Figure 1Learning curve by grid search for SAM ρ determination. (A) accuracy score (B) loss scoreIn Epoch300, the convergence was good when ρ was 0.01 or 0.025. In the comparison of ρ = 0.025 and 0.01 in Loss, 0.025 was more stable.
Figure 2Learning curve in each CNN model.
Comparison of optimizer SAM and SGD with and without a learning rate scheduler.
| Optimizer | Learning rate | Accuracy | Precision | Recall | F1 score | AUC |
|---|---|---|---|---|---|---|
| SD | SD | SD | SD | SD | ||
| 95% CI | 95% CI | 95% CI | 95% CI | 95% CI | ||
| w/o scheduler | 0.8879 | 0.8518 | 0.6501 | 0.6937 | 0.9020 | |
| 0.0016 | 0.0099 | 0.0089 | 0.0095 | 0.0075 | ||
| 0.8873–0.8885 | 0.8482–0.8553 | 0.6469–0.6533 | 0.6903–0.6971 | 0.8993–0.9047 | ||
| With scheduler | 0.8970 | 0.8049 | 0.7581 | 0.7780 | 0.9098 | |
| 0.0026 | 0.0066 | 0.0064 | 0.0057 | 0.0030 | ||
| 0.8961–0.8979 | 0.8026–0.8073 | 0.7558–0.7604 | 0.7759–0.7800 | 0.9087–0.9108 | ||
| w/o scheduler | 0.8870 | 0.8529 | 0.6438 | 0.6872 | 0.8766 | |
| 0.0004 | 0.0024 | 0.0017 | 0.0019 | 0.0069 | ||
| 0.8868–0.8871 | 0.8521–0.8538 | 0.6432–0.6444 | 0.6875–0.6879 | 0.8741–0.8791 | ||
| With scheduler | 0.9016 | 0.8534 | 0.7139 | 0.7578 | 0.9328 | |
| 0.0018 | 0.0072 | 0.0124 | 0.0098 | 0.0016 | ||
| 0.9010–0.9023 | 0.8509–0.8560 | 0.7094–0.7183 | 0.7543–0.7614 | 0.9322–0.9334 | ||
SD, standard deviation; 95% CI, 95% confidence interval; AUC, Area under the ROC curve.
Statistical evaluation of optimizers SAM and SGD with and without a learning rate scheduler.
| Performance metrics | Model B | Model A | A–B | Effect size | |
|---|---|---|---|---|---|
| Accuracy | w/o LRS | With LRS | 0.0091 | < .0001 | 4.143 |
| Precision | w/o LRS | With LRS | − 0.0468 | < .0001 | 5.463 |
| Recall | w/o LRS | With LRS | 0.1080 | < .0001 | 13.721 |
| F1 score | w/o LRS | With LRS | 0.0843 | < .0001 | 10.609 |
| AUC | w/o LRS | With LRS | 0.0078 | < .0001 | 1.348 |
| Accuracy | w/o LRS | With LRS | 0.0147 | < .0001 | 11.226 |
| Precision | w/o LRS | With LRS | 0.0005 | 0.733 | 0.090 |
| Recall | w/o LRS | With LRS | 0.0701 | < .0001 | 7.832 |
| F1 score | w/o LRS | With LRS | 0.0706 | < .0001 | 9.894 |
| AUC | w/o LRS | With LRS | 0.0562 | < .0001 | 10.997 |
| Accuracy | SDG | SAM | 0.0046 | < .0001 | 2.047 |
| Precision | SDG | SAM | 0.0485 | < .0001 | 6.924 |
| Recall | SDG | SAM | − 0.0442 | < .0001 | 4.432 |
| F1 score | SDG | SAM | − 0.0201 | < .0001 | 2.476 |
| AUC | SDG | SAM | 0.0230 | < .0001 | 9.529 |
AUC, area under the ROC curve; LRS, learning rate scheduler.
Figure 3Visualization of regions of interest for CNN classification in oral exfoliative cytopathology classification. In the heat map visualization, the warmer the color, the greater the effect on label classification.
Figure 4Automatic division of WSI using OpenSlide. Images can be acquired from different depth levels. In this study, images with a depth of 14th level were used.
Label data distribution.
| Class | Number of images | Description (From papanicolaou, 1954) |
|---|---|---|
| ClassI | 2210 | Absence of atypical or abnormal cells |
| ClassII | 2903 | Atypical cytology, but no evidence for malignancy |
| ClassIII | 225 | Cytology suggestive of, but not conclusive for, malignancy |
| ClassIV | 416 | Cytology strongly suggestive of malignancy |
| ClassV | 240 | Cytology conclusive for malignancy |
| Total | 5994 | |
| Unclassifiable images | 3720 | |
Figure 5Overall flow of deep-learning classification model research of oral exfoliative cytopathology.