| Literature DB >> 35741267 |
Guan-Hua Huang1, Qi-Jia Fu1, Ming-Zhang Gu1, Nan-Han Lu2,3,4, Kuo-Ying Liu5, Tai-Been Chen1,4.
Abstract
Chest X-ray (CXR) is widely used to diagnose conditions affecting the chest, its contents, and its nearby structures. In this study, we used a private data set containing 1630 CXR images with disease labels; most of the images were disease-free, but the others contained multiple sites of abnormalities. Here, we used deep convolutional neural network (CNN) models to extract feature representations and to identify possible diseases in these images. We also used transfer learning combined with large open-source image data sets to resolve the problems of insufficient training data and optimize the classification model. The effects of different approaches of reusing pretrained weights (model finetuning and layer transfer), source data sets of different sizes and similarity levels to the target data (ImageNet, ChestX-ray, and CheXpert), methods integrating source data sets into transfer learning (initiating, concatenating, and co-training), and backbone CNN models (ResNet50 and DenseNet121) on transfer learning were also assessed. The results demonstrated that transfer learning applied with the model finetuning approach typically afforded better prediction models. When only one source data set was adopted, ChestX-ray performed better than CheXpert; however, after ImageNet initials were attached, CheXpert performed better. ResNet50 performed better in initiating transfer learning, whereas DenseNet121 performed better in concatenating and co-training transfer learning. Transfer learning with multiple source data sets was preferable to that with a source data set. Overall, transfer learning can further enhance prediction capabilities and reduce computing costs for CXR images.Entities:
Keywords: convolutional neural network; deep learning; source data set; supervised classification
Year: 2022 PMID: 35741267 PMCID: PMC9222116 DOI: 10.3390/diagnostics12061457
Source DB: PubMed Journal: Diagnostics (Basel) ISSN: 2075-4418
Characteristics of the included data sets.
| Category | Name | Label Category | Size | Feature |
|---|---|---|---|---|
| Source data | ImageNet | 20,000 + | 14 million + | Large and diverse |
| Source data | CheXpert | 14 1 | 224,316 | Similar to the target data |
| Source data | ChestX-ray | 14 2 | 112,010 4 | Similar to the target data |
| Target data | E-Da chest X-ray | 8 3 | 1630 | Small but important |
1 Thirteen common thoracic disease labels and a “no finding” label (indicating the absence of any disease). 2 Thirteen disease labels and a normal label. 3 Seven disease labels and a normal label. 4 One hundred and ten images labeled as “hernia” in the original dataset were discarded in this analysis due to the small sample size.
Numbers and labels of images in the target data set.
| Category | Sample Size | Subcategory | Sample Size |
|---|---|---|---|
| normal | 1212 | normal | 1212 |
| aortic sclerosis/calcification | 90 | aortic arch atherosclerotic plaque | 90 |
| arterial curvature | 93 | tortuous aorta/thoracic aortic ectasia | 93 |
| abnormal lung fields | 33 | shadows of pulmonary nodules | 13 |
| consolidations or lung cavities in the upper lobes | 5 | ||
| pulmonary fibrosis | 15 | ||
| increased lung patterns | 153 | atelectasis/focal consolidation | 109 |
| enlarged hilar shadow | 44 | ||
| spinal lesions | 148 | degenerative joint disease of the thoracic spine | 75 |
| scoliosis | 73 | ||
| cardiomegaly | 41 | cardiomegaly | 41 |
| intercostal pleural thickening | 36 | intercostal pleural thickening | 36 |
Figure 1Flow chart of our deep transfer learning approach for the multilabel classification of the chest X-ray images.
Figure 2Flow chart of concatenating transfer learning.
Figure 3Flow chart of co-training transfer learning.
Summary of experimental configurations and parameter settings.
| Configuration | Setting |
|---|---|
| Target data set | CXR images from the E-Da Hospital, Taiwan |
| Target task | Multilabel classification of CXR images |
| Loss function | Weighted binary cross entropy |
| Image augmentation | The ImageDataGenerator function in the Keras Python library for online augmentation |
| Deep CNN modeling | |
| Backbone architecture | The Python package Keras to implement 50-layer ResNet and 121-layer DenseNet |
| Optimizer | Adam |
| Mini-batch size | 16 |
| Epoch number | 30 |
| Learning rate | Start from 0.0001 and is divided by 10 when the validation loss does not decrease in 10 epochs |
| Growth rate in DenseNet | 12 |
| Transfer learning | |
| Source data set | ImageNet, ChestX-ray, and CheXpert |
| Reuse pretrained weights | Model finetuning and layer transfer |
| Combine source data sets | Initiating, concatenating, and co-training |
| Evaluation | |
| Data splitting | The Python package MultilabelStratifiedKFold for stratified |
| Metrics | Accuracy, AUC, and AP |
Mean test accuracies and AUCs of different approaches of reusing pretrained weights in ResNet50 1.
| Pretrained Weight | ||||||||
|---|---|---|---|---|---|---|---|---|
| ImageNet with Random Initials | ChestX-ray with Random Initials | ChestX-ray with ImageNet Initials | Random Initials 2 | |||||
| Method | AC 4 | AUC | AC | AUC | AC | AUC | AC | AUC |
| Layer transfer 3 | 0.775 | 0.765 | ||||||
| RB_1 |
| 0.487 | 0.557 |
| 0.461 |
| ||
| RB_2 | 0.828 | 0.488 | 0.748 | 0.514 | 0.749 | 0.511 | ||
| RB_3 | 0.814 | 0.469 |
| 0.496 |
| 0.505 | ||
| RB_4 | 0.771 |
| 0.512 | 0.498 | 0.646 | 0.499 | ||
| Model finetuning |
|
|
|
|
|
| ||
1 Bold numbers indicate the top two approaches in each metric. 2 The model without transfer learning. 3 RB_1 = Pretrained weights on the first 10 layers, which correspond to the end of the first residual block; RB_2 = Pretrained weights on the first 22 layers, which correspond to the end of the second residual block; RB_3 = Pretrained weights on the first 40 layers, which correspond to the end of the third residual block; RB_4 = Pretrained weights on the first 49 layers, which correspond to the end of the fourth residual block. 4 AC = Accuracy.
Training and test performance of various transfer learning approaches 1.
| Backbone Model | Transfer Learning Approach 2 | Mean Training Accuracy | Mean Test AUC | Mean Test AP |
|---|---|---|---|---|
| Initiating transfer learning | ||||
| ResNet50 | IR |
| 0.796 |
|
| ResNet50 | CR | 0.879 |
| 0.209 |
| ResNet50 | CI | 0.868 |
| 0.206 |
| ResNet50 | XR | 0.819 | 0.806 | 0.191 |
| ResNet50 | XI | 0.852 |
|
|
| DenseNet121 | IR | 0.916 | 0.803 | 0.204 |
| DenseNet121 | CR | 0.807 | 0.800 | 0.171 |
| DenseNet121 | CI | 0.784 | 0.779 | 0.179 |
| DenseNet121 | XR | 0.799 | 0.781 | 0.169 |
| DenseNet121 | XI | 0.864 | 0.826 |
|
|
| ||||
| ResNet50 |
|
| 0.780 | 0.207 |
| ResNet50 |
| 0.930 | 0.776 | 0.190 |
| DenseNet121 |
|
| 0.802 | 0.210 |
| DenseNet121 |
| 0.914 | 0.813 | 0.210 |
|
| ||||
| ResNet50 |
| 0.855 | 0.790 | 0.191 |
| DenseNet121 |
| 0.775 | 0.826 | 0.210 |
1 Bold numbers indicate the top three approaches in each metric. 2 IR = ImageNet pretraining with random initials, CR = ChestX-ray pretraining with random initials, CI = ChestX-ray pretraining with ImageNet pretraining initials, XR = CheXpert pretraining with random initials, XI = CheXpert pretraining with ImageNet pretraining initials, = Concatenating ImageNet + ChestX-ray, = Concatenating ImageNet + CheXpert, = Co-training ChestX-ray + CheXpert.