| Literature DB >> 36174073 |
Ahmed Saihood1,2, Hossein Karshenas1, Ahmad Reza Naghsh Nilchi1.
Abstract
Lung cancer is a serious threat to human health, with millions dying because of its late diagnosis. The computerized tomography (CT) scan of the chest is an efficient method for early detection and classification of lung nodules. The requirement for high accuracy in analyzing CT scan images is a significant challenge in detecting and classifying lung cancer. In this paper, a new deep fusion structure based on the long short-term memory (LSTM) has been introduced, which is applied to the texture features computed from lung nodules through new volumetric grey-level-co-occurrence-matrices (GLCMs), classifying the nodules into benign, malignant, and ambiguous. Also, an improved Otsu segmentation method combined with the water strider optimization algorithm (WSA) is proposed to detect the lung nodules. WSA-Otsu thresholding can overcome the fixed thresholds and time requirement restrictions in previous thresholding methods. Extended experiments are used to assess this fusion structure by considering 2D-GLCM based on 2D-slices and approximating the proposed 3D-GLCM computations based on volumetric 2.5D-GLCMs. The proposed methods are trained and assessed through the LIDC-IDRI dataset. The accuracy, sensitivity, and specificity obtained for 2D-GLCM fusion are 94.4%, 91.6%, and 95.8%, respectively. For 2.5D-GLCM fusion, the accuracy, sensitivity, and specificity are 97.33%, 96%, and 98%, respectively. For 3D-GLCM, the accuracy, sensitivity, and specificity of the proposed fusion structure reached 98.7%, 98%, and 99%, respectively, outperforming most state-of-the-art counterparts. The results and analysis also indicate that the WSA-Otsu method requires a shorter execution time and yields a more accurate thresholding process.Entities:
Mesh:
Year: 2022 PMID: 36174073 PMCID: PMC9521911 DOI: 10.1371/journal.pone.0274516
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.752
Fig 1The four stages of lung cancer development.
Fig 2The constituent steps of the proposed method.
Fig 3The 3D-GLCMs are computed from VSs of nodule volume and fed to a recurrent neural network.
Fig 4In 2.5D-GLCM mode the GLCMs computed for each slice are fed to the recurrent neural network.
The structure of the proposed LSTM-based fusion method.
f refers to the number of elements in the extracted features.
| Parameters | 2D-GLCM-LSTM | 2.5D-GLCM-LSTM | 3D-GLCM-LSTM |
|---|---|---|---|
| Input shape | 8 × | ||
| Layer 1 | LSTM, 128 units, | LSTM, 128 units, | LSTM, 128, units, |
| Layer 2 | LSTM, 128, units, | LSTM, 128, units, | LSTM, 128, units, |
| Layer 3 | Dense, 3 units, | Dense, 32 units, | Dense, 32 units, |
| Layer 4 | × | Dense, 3 units, | Dense, 3 units, |
The comparison of WSA-Otsu with counterpart methods.
| Methods | TP | TN | FP | FN | F1-Score |
|---|---|---|---|---|---|
| Standard Otsu method [ | 95 | 38 | 5 | 17 | 89.6% |
| Otsu based Darwinian particle swarm optimization (DPSO) [ | 89 | 49 | 7 | 10 | 91.3% |
| Statistical and Shape-Based Features Lung nodules detection [ | 101 | 32 | 6 | 16 | 90.2% |
| Otsu based exchange market algorithm (EMA) [ | 93 | 48 | 11 | 3 | 93.0% |
| Adaptive Particle Swarm Optimization– Gaussian mixture model (APSO-GMM) [ | 109 | 33 | 8 | 5 | 94.4% |
| FFNN + combination of cuckoo and PSO [ | 112 | 33 | 6 | 4 | 95.7% |
|
|
|
|
|
|
|
Fig 5The convergence curves of Otsu method in terms of fitness values at five different levels.
Fig 6The thresholding time of the proposed method vs. other counterparts.
Fig 7ROC curves for three fusion models and their corresponding macro-averaged AUC values.
Fig 8ROC curves for three fusion models and their corresponding micro-averaged AUC values.
A comparison between the proposed 2D and 2.5D GLCM fusion methods and other deep CNN-based methods regarding the accuracy, sensitivity, and specificity metrics.
| Method | Accuracy ± | Sensitivity ± | Specificity ± |
|---|---|---|---|
| MobileNetV2 [ | 82.97 ±1.3 | 69.37 ±.8 | 92.12 ±1.1 |
| EfficientNet-B5 [ | 88.77±1.7 | 94.59±0.2 | 84.85±0.5 |
| ResNet50 [ | 86.23±2.1 | 98.2±0.2 | 78.18±1.02 |
| Xception [ | 92.39±1.2 | 93.69±1.55 | 91.52±0.9 |
| NASNetMobile [ | 87.68±0.22 | 74.77±0.6 | 96.36±0.56 |
| Raw CTs-based 3D-CNN [ | 97.17±0.25 | 87±0.57 | 94±0.11 |
| LSTM-based deep fusion of 2D-GLCMs | 94.4±0.7 | 91.6±0.32 | 95.6±0.58 |
|
|
|
|
|
A comparison between the proposed fusion methods for different modes of GLCM computation and the recently proposed classification methods regarding the accuracy, sensitivity, and specificity metrics.
| Research | Method | Accuracy | Sensitivity | Specificity |
|---|---|---|---|---|
| [ | Linear Discriminant analysis | × | 70% | 78% |
| [ | MTANN | × | 80.30% | 83.5% |
| [ | FLD classifier | × | 82.66 | 84.6% |
| [ | a neural classifier | × | 87.50% | 89.3% |
| [ | cylindrical filters and SVM | × | 80% | 88.6% |
| [ | Hierarchical Vector Quantization and SVM | × | 82.70% | 86.3% |
| [ | GLMR classifier | × | 92.91% | 89.3% |
| [ | LDA classifier and optimal thresholding | 84% | 97.14% | 80.2% |
| [ | backpropagation network | 90.70% | × | 87.7% |
| [ | fuzzy inference method | 94.12% | × | 96% |
| [ | texture and learned distance metrics and TSCBIR classifier | 91% | × | 81.1% |
| [ | probabilistic neural network | 92% | 95% | 90% |
| [ | multilayer feed-forward neural network with supervised learning method as a classifier | 95% | 100% | 93.5% |
| [ | double-path convolutional neural network (DPCNN) | 70.4% | × | 77.8% |
| [ | local energy-based shape histograms + Adaboost | 84.6% | 71.4% | 100% |
| [ | 3D-CNN-based scans. | 83.33% | × | 94% |
| [ | GLCM based CNN-RNN | 76% | × | 84.5% |
| [ | 3D-GLCM-CNN | 93% | 90% | 89% |
| [ | 3D-convolution-LSTM | 97.2% | 98.2% | 92.6% |
| [ | CNN-LSTM | 97% | × | 91.7% |
| [ | Fusion of clinical and CT-scan-based features | 93.6% | 91.9% | 95.6% |
| This study |
| 95.6±0.58% | ||
| This study |
| 98±0.66% | ||
| This study |
| 99±0.98% |
Fig 9ROC curves for baseline and proposed models, and their corresponding micro-averaged AUC values.
Fig 10ROC curves for baseline and proposed models, and their corresponding macro-averaged AUC values.
The classification results in terms of micro-averaged AUCs trained on the LIDC-IDRI dataset and tested on LIDC-IDRI and LUNGx datasets for the proposed and baseline models.
| models | AUCs± | |
|---|---|---|
| LUNGx dataset | LIDC-IDRI dataset | |
| CNN-LSTM fusion | 51.3±7.1 | 90.5±1.9 |
| 3D GLCM based CNN fusion | 56.05±5.1 | 89.7 ±2.4 |
| 2.5D GLCM based LSTM fusion | 61.3±4.9 | 92.9±2.6 |
| 3D GLCM based LSTM fusion | 70.02±3.5 | 93.1±1.8 |
The T-test p-values obtained comparing the average AUC of the LSTM-based deep fusion of 3D-GLCMs with other baseline models.
| models | |
|---|---|
| CNN-LSTM fusion | 0.0017 |
| 3D GLCM based CNN fusion | 0.000018 |
| 2.5D GLCM based LSTM fusion | 0.02 |
Results of the ablation study for the type of features presented to the LSTM-based deep fusion model.
| Features used | Accuracy | Sensitivity | Specificity |
|---|---|---|---|
| 3D-GLCMs stacked for each VS | 98.7% | 98% | 99% |
| Texture feature descriptors | 91.3% | 89.3% | 90.3% |
| 3D-GLCMs stacked for each direction | 92.7% | 91.2% | 88.3% |
Fig 11The training time of different models for five epochs.
Fig 12The test time of different models.