| Literature DB >> 35978109 |
Shaoli Wang1, Yingying Chen1, Siying Chen1, Qionglei Zhong2, Kaiyan Zhang3.
Abstract
Laryngeal disease classification is a relatively hard task in medical image processing resulting from its complex structures and varying viewpoints in data collection. Some existing methods try to tackle this task via the convolutional neural network, but they more or less ignore the intrinsic difficulty differences among different input samples and suffer from high training complexity. In order to better resolve these problems, an end-to-end Hierarchical Dynamic Convolutional Network (HDCNet) is proposed, which can dynamically process the input samples based on their difficulty. For the easy-classified samples, the HDCNet processes them with a smaller resolution and a relatively small network, while the difficult samples are passed to a large network with a larger resolution for more accurate classification results. Furthermore, a Feature Reuse Module (FRM) is designed to transfer the features learned by the small network to the corresponding block in the deep network to enhance the overall performance of some rather complicated samples. To validate the effectiveness of the proposed HDCNet, comprehensive experiments are conducted on the public available laryngeal disease classification dataset and HDCNet provides superior performances compared with other current state-of-the-art methods.Entities:
Mesh:
Year: 2022 PMID: 35978109 PMCID: PMC9385650 DOI: 10.1038/s41598-022-18217-5
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.996
Figure 1The overall pipeline of the proposed HDCNet.
Figure 2The overall pipeline of the proposed Feature Reuse Module.
The performance evaluation of different methods on the AUC and overall accuracy among all classes, the best results are highlighted in bold.
| Methods | Edema | Cancer | Granuloma | Normal | Leukoplakia | Cyst | Nodules | Polyps | average AUC | Accuracy |
|---|---|---|---|---|---|---|---|---|---|---|
| CheXNet[ | 0.798 | 0.822 | 0.979 | 0.900 | 0.876 | 0.685 | 0.825 | 0.853 | 0.843 | 71% |
| AG-CNN[ | 0.805 | 0.879 | 0.972 | 0.895 | 0.896 | 0.658 | 0.828 | 0.838 | 0.847 | 71% |
| Xiong et al.[ | 0.857 | 0.866 | 0.978 | 0.911 | 0.894 | 0.705 | 0.857 | 0.886 | 0.870 | 71% |
| Yin et al | 0.900 | 0.936 | 0.965 | 0.878 | 0.853 | 0.849 | 0.886 | 0.871 | 0.893 | 73% |
| HDCNet | 0.921 | 0.953 | 0.835 |
Figure 3Comparison between different methods.
Comparison of different methods with the proposed HDCNet according to the Accuracy and FLOPs.
| Model (resolution) | Accuracy (FLOPs) | |
|---|---|---|
| Single model | MobileNetv3_large_w1 (224) | 52.52% (0.12G) |
| EfficientNet_b0 (224) | 54.17% (0.21G) | |
| ResNet18 (224) | 70.95% (0.91G) | |
| ResNet34 (336) | 73.58% (4.21G) | |
| ResNet50 (336) | 74.15% (4.73G) | |
| HDCNet (combination) | ResNet18 (224) + ResNet34 (336) | 75.27% (1.74G) |
| ResNet18 (224) + ResNet50 (336) | 75.57% (1.87G) | |
| ResNet34 (224) + ResNet50 (336) | 75.90% (2.56G) |
Figure 4(a) and (b) are input images that cannot be classified correctly by ResNet18 but can be correctly diagnosed by the proposed HDCNet.
Figure 5The performance with respect to different thresholds . The accuracy is used as the metric for selecting the best .