| Literature DB >> 29792208 |
Changmiao Wang1,2, Ahmed Elazab3,4, Fucang Jia1, Jianhuang Wu1, Qingmao Hu5,6.
Abstract
OBJECTIVE: In this paper, we aim to investigate the effect of computer-aided triage system, which is implemented for the health checkup of lung lesions involving tens of thousands of chest X-rays (CXRs) that are required for diagnosis. Therefore, high accuracy of diagnosis by an automated system can reduce the radiologist's workload on scrutinizing the medical images.Entities:
Keywords: Autoencoder; Chest screening; Computer aided diagnosis; Deep learning; Receiver operating characteristic
Mesh:
Year: 2018 PMID: 29792208 PMCID: PMC5966927 DOI: 10.1186/s12938-018-0496-2
Source DB: PubMed Journal: Biomed Eng Online ISSN: 1475-925X Impact factor: 2.819
Number of positive cases and negative cases used in training and validation sets
| Training set | Validation set | ||
|---|---|---|---|
| Positive | Negative | Positive | Negative |
| 418 | 1318 | 179 | 565 |
Fig. 1Workflow diagram of our proposed method
Fig. 2Segmentation of lung from public database based on U-net
Fig. 3Flow chart of Autoencoder
Network architecture of CSDAE
| Layer (type) | Output shape | Param# |
|---|---|---|
| Input_1(InputLayer) | (None, 512, 512, 1) | 0 |
| Conv2d_1(Conv2D) | (None, 512, 512, 16) | 160 |
| Max_pooling2d_1(MaxPooling2D) | (None, 256, 256, 16) | 0 |
| Conv2d_2(Conv2D) | (None, 256, 256, 8) | 1160 |
| Max_pooling2d_2(MaxPooling2D) | (None, 128, 128, 8) | 0 |
| Conv2d_3(Conv2D) | (None, 128, 128, 8) | 584 |
| Max_pooling2d_3(MaxPooling2D) | (None, 64, 64, 8) | 0 |
| Conv2d_4(Conv2D) | (None, 64, 64, 8) | 584 |
| Up_sampling2d_1(UpSampling2D) | (None, 128, 128, 8) | 0 |
| Conv2d_5(Conv2D) | (None, 128, 128, 8) | 584 |
| Up_sampling2d_2(UpSampling2D) | (None, 256, 256, 8) | 0 |
| Conv2d_6(Conv2D) | (None, 256, 256, 8) | 1168 |
| Up_sampling2d_3(UpSampling2D) | (None, 512, 512, 16) | 0 |
| Conv2d_7(Conv2D) | (None, 512, 512, 1) | 145 |
Fig. 4Loss over the epochs on the AE and CSDAE. a AE with noise factor 0.01, b CSDAE with noise factor 0.01, c AE with noise factor 0.05, and d CSDAE with noise factor 0.05
Network architecture of proposed image classification
| Layer (type) | Output shape | Param# |
|---|---|---|
| Input_4(InputLayer) | (None, 512) | 0 |
| Batch_normalization_4(Batch) | (None, 512) | 2048 |
| Dropout_4(Dropout) | (None, 512) | 0 |
| Dense_4(Dense) | (None, 512) | 262,656 |
| Batch_normalization_5(Batch) | (None, 512) | 2048 |
| Dropout_5(Dropout) | (None, 512) | 0 |
| Dense_5(Dense) | (None, 512) | 262,656 |
| Batch_normalization_6(Batch) | (None, 512) | 2048 |
| Dropout_6(Dropout) | (None, 512) | 0 |
| Dense_6(Dense) | (None, 1) | 513 |
Comparisons of the mean and std of MSE between AE and CSDAE
| Noise factor | Train data MSE (mean ± std) | Test data MSE (mean ± std) | Test noised data MSE (mean ± std) | |
|---|---|---|---|---|
| AE | 0.01 | 0.000983 ± 0.000304 | 0.001088 ± 0.0004 | 0.001112 ± 0.000405 |
| CSDAE | 0.01 | 0.000829 ± 0.000282 | 0.000923 ± 0.000369 | 0.000922 ± 0.000368 |
| AE | 0.05 | 0.001039 ± 0.000334 | 0.001146 ± 0.000427 | 0.001255 ± 0.000445 |
| CSDAE | 0.05 | 0.001003 ± 0.000334 | 0.001103 ± 0.000421 | 0.001174 ± 0.00045 |
Fig. 5Original lung image and the reconstructed lung image from AE: upper row is the original image while the bottom row is the reconstruction image
Fig. 6Original lung image and reconstructed lung image from CSDAE. a Results of CSDAE with noise factor 0.01: the upper row is the original image, middle row is the image with noise, and bottom row is the reconstructed image. b Results of CSDAE with noise factor 0.05: the upper row is the original image, middle row is the image with noise, and bottom row is the reconstructed image
Fig. 7Precision for different threshold in CSDAE (a) with noise factor 0.01 and (b) noise factor 0.05
Performance results based on test data using four classifiers: KNN, logistic regression, SVM, random forest
| Classifier | Data augmentation | Precision | Recall | F1 | AUC |
|---|---|---|---|---|---|
| KNN | Without augmentation | 0.73 | 0.58 | 0.64 | 0.753 |
| 0.87 | 0.93 | 0.9 | |||
| SMOTE | 0.44 | 0.84 | 0.58 | 0.752 | |
| 0.93 | 0.67 | 0.78 | |||
| Positive augmentation | 0.57 | 0.68 | 0.62 | 0.758 | |
| 0.89 | 0.68 | 0.87 | |||
| 4× augmentation | 0.57 | 0.84 | 0.62 | 0.758 | |
| 0.89 | 0.89 | 0.87 | |||
| Logistic | Without augmentation | 0.76 | 0.54 | 0.63 | 0.744 |
| 0.87 | 0.95 | 0.91 | |||
| SMOTE | 0.61 | 0.82 | 0.64 | 0.79 | |
| 0.92 | 0.76 | 0.84 | |||
| Positive augmentation | 0.57 | 0.66 | 0.61 | 0.754 | |
| 0.89 | 0.84 | 0.86 | |||
| 4× augmentation | 0.63 | 0.72 | 0.67 | 0.792 | |
| 0.91 | 0.87 | 0.89 | |||
| SVM | Without augmentation | 0.63 | 0.73 | 0.68 | 0.798 |
| 0.91 | 0.86 | 0.89 | |||
| SMOTE | 0.63 | 0.77 | 0.69 | 0.813 | |
| 0.92 | 0.86 | 0.89 | |||
| Positive augmentation | 0.59 | 0.64 | 0.61 | 0.75 | |
| 0.88 | 0.86 | 0.87 | |||
| 4× augmentation | 0.61 | 0.71 | 0.66 | 0.784 | |
| 0.9 | 0.86 | 0.88 | |||
| Random forest | Without augmentation | 0.69 | 0.38 | 0.49 | 0.663 |
| 0.83 | 0.95 | 0.88 | |||
| SMOTE | 0.62 | 0.5 | 0.56 | 0.704 | |
| 0.85 | 0.9 | 0.88 | |||
| Positive augmentation | 0.54 | 0.63 | 0.58 | 0.73 | |
| 0.88 | 0.83 | 0.85 | |||
| 4× augmentation | 0.59 | 0.54 | 0.56 | 0.71 | |
| 0.86 | 0.88 | 0.87 |
Comparison of accuracy, recall, F1 score, and AUC of the methods on test data by the deep network based on four data augmentation methods
| Loss function | Data augmentation | Precision | Recall | F1 | AUC |
|---|---|---|---|---|---|
| Cross entropy | Without augmentation | 0.76 | 0.72 | 0.74 | 0.821 |
| 0.91 | 0.93 | 0.92 | |||
| SMOTE | 0.58 | 0.82 | 0.68 | 0.813 | |
| 0.93 | 0.81 | 0.87 | |||
| Positive augmentation | 0.67 | 0.75 | 0.81 | 0.815 | |
| 0.92 | 0.88 | 0.90 | |||
| 4× augmentation | 0.69 | 0.77 | 0.72 | 0.827 | |
| 0.92 | 0.89 | 0.91 | |||
| Focal loss | Without augmentation | 0.76 | 0.73 | 0.74 | 0.828 |
| 0.91 | 0.93 | 0.92 | |||
| SMOTE | 0.52 | 0.82 | 0.64 | 0.79 | |
| 0.93 | 0.76 | 0.84 | |||
| Positive augmentation | 0.74 | 0.72 | 0.73 | 0.817 | |
| 0.91 | 0.92 | 0.91 | |||
| 4× augmentation | 0.7 | 0.74 | 0.72 | 0.821 | |
| 0.9 | 0.9 | 0.91 |
Results of the performance of our network method
Fig. 8Recalls for different thresholds in CSDAE (a) with noise factor 0.01 and (b) noise factor 0.05
The results of the performance of our method with max AUC