| Literature DB >> 36097497 |
Debnath Bhattacharyya1, N Thirupathi Rao2, Eali Stephen Neal Joshua2, Yu-Chen Hu3.
Abstract
Lung nodules are abnormal growths and lesions may exist. Both lungs may have nodules. Most lung nodules are harmless (not cancerous/malignant). Pulmonary nodules are rare in lung cancer. X-rays and CT scans identify the lung nodules. Doctors may term the growth a lung spot, coin lesion, or shadow. It is necessary to obtain properly computed tomography (CT) scans of the lungs to get an accurate diagnosis and a good estimate of the severity of lung cancer. This study aims to design and evaluate a deep learning (DL) algorithm for identifying pulmonary nodules (PNs) using the LUNA-16 dataset and examine the prevalence of PNs using DB-Net. The paper states that a new, resource-efficient deep learning architecture is called for, and it has been given the name of DB-NET. When a physician orders a CT scan, they need to employ an accurate and efficient lung nodule segmentation method because they need to detect lung cancer at an early stage. However, segmentation of lung nodules is a difficult task because of the nodules' characteristics on the CT image as well as the nodules' concealed shape, visual quality, and context. The DB-NET model architecture is presented as a resource-efficient deep learning solution for handling the challenge at hand in this paper. Furthermore, it incorporates the Mish nonlinearity function and the mask class weights to improve segmentation effectiveness. In addition to the LUNA-16 dataset, which contained 1200 lung nodules collected during the LUNA-16 test, the LUNA-16 dataset was extensively used to train and assess the proposed model. The DB-NET architecture surpasses the existing U-NET model by a dice coefficient index of 88.89%, and it also achieves a similar level of accuracy to that of human experts.Entities:
Keywords: Bidirectional feature extraction; Computer-aided diagnosis; Convolutional neural network; Deep learning; Lung cancer
Year: 2022 PMID: 36097497 PMCID: PMC9453728 DOI: 10.1007/s00371-022-02657-1
Source DB: PubMed Journal: Vis Comput ISSN: 0178-2789 Impact factor: 2.835
Related research and gap identification
| References | Research objective | Segmentation technique | Research Gap | Split (%) | Accuracy (%) |
|---|---|---|---|---|---|
| [ | Dilated multi-residual blocks network based on U-NET for biomedical image segmentation | The segmented image of DC-U-NET is closer to the ground truth than Otsu, and the region is growing | The position of the CT scan is not accurate. The findings of this study hold up to minimal data | 70:30 | 85.97 |
| [ | Automatic detect lung node with deep learning in segmentation and imbalance data labeling | Semantic Segmentation | The position of the CT scan is not accurate and the label goes away from the point of intersection. The findings of this study hold up to minimal data | 60:40 | 85.57 |
| [ | A deep learning model to automate skeletal muscle area measurement on computed tomography images | Ensemble Learning with Semantic Segmentation | This model failed to work on the high-quality digital scans. When high-quality scans were given as input failed to give better accuracy | 70:30 | 84.23 |
| [ | Multi-level Seg-Unet model with global and patch-based X-ray images for knee bone tumor detection | Patch-based X-ray images for knee bone tumor detection | The mean accuracy of the model was not benchmarked. Thus, we cannot rely on the model performance | 70:30 | 84.81 |
| [ | Lung cancer and granuloma identification using a deep learning model to extract 3-dimensional radionics features in CT imaging | Tumor segmentation and radiomics feature extraction of the region of interest using gradient boosting | The prediction of the model was not standard. Working efficiently when less amount of data is fed to the classifier | 60:40 | 83.22 |
| [ | ResBCDU-NET: A deep learning framework for lung CT image segmentation | Bidirectional Convolutional Long Short-term Memory is used as an advanced integrator module | This model failed in the identification of similar image densities. As a result, the model performance was not up to the mark | 70:30 | 83.74 |
| [ | Automated lung segmentation on chest computed tomography images with extensive lung parenchymal abnormalities using a deep neural network | 2D-UNET with three-dimension feature extraction | This model failed when the image had a group of voxels when the data were large | 60:40 | 84.56 |
| [ | Conventional filtering versus U-NET-based models for pulmonary nodule segmentation in CT images | Semantic segmentation with Seg-U-NET | This model failed in time computational cost and memory computation | 70:30 | 83.65 |
| [ | Efficacy evaluation of 2D, 3D U-NET semantic segmentation and atlas-based segmentation of normal lungs excluding the trachea and main bronchi | Semantic segmentation | This model requires high-performance computation equipment | 70:30 | 82.63 |
| [ | Semantic segmentation and detection of mediastinal lymph nodes and anatomical structures in CT data for lung cancer staging | Pixelwise Bio-Medical Image semantic segmentation and Instance Segmentation | This model failed to identify the instances detected and represented by a 3D pixelwise mask, bounding volume, and centroid position | 70:30 | 81.54 |
| [ | Fully automated lung lobe segmentation in volumetric chest ct with 3D U-NET: Validation with intra- and extra-datasets | Deep convolutional neural network for the image segmentation | The model failed to meet gold standards when it came to performance the accuracy | 70:30 | 82.65 |
| [ | 3D ResNetwork-I for automatic detection of Lung Nodules in CT scans | Deep convolutional neural network for the image segmentation | Global and local features in the CT scan images could not segment properly | 70:30 | 82.36 |
| [ | Medical Image segmentation using Encoding and Decoding using the Deep learning approaches | X-Shaped convolutional neural network | In encoding and decoding model was weakened during the insufficient data | 70:30 | 81.67 |
| [ | Segmentation and detection on the unsupervised data | Image synthesis and image anomalies using U-NET | They drastically failed to identify the abnormal samples | 65:35 | 85.69 |
Fig. 1The basic U-NET architecture [31]
Fig. 2The graphical representation of Mish activation
Fig. 4An example of image augmentation after flipping and rotating
The features of the LUNA16 dataset in the standard deviation format
| Characteristics | Training set | Testing set |
|---|---|---|
| Malignancy | 2.96 ± 0.96 | 3.04 ± 1.02 |
| Speculation | 1.61 ± 0.80 | 1.66 ± 0.88 |
| Subtlety | 3.92 ± 0.84 | 4.08 ± 0.79 |
| Lobulation | 1.74 ± 0.74 | 1.83 ± 0.81 |
| Diameter in mm | 8.14 ± 4.59 | 9.08 ± 5.25 |
| Margin | 4.04 ± 0.84 | 4.07 ± 0.78 |
Fig. 3The proposed architecture for lung cancer segmentation, where the down-sampling and up-sampling were stopped between the convolutional neural network
The proposed architecture layers and their respective parameters, activations, and output
| Layers | Parameters | Activation | Output |
|---|---|---|---|
| Convolution 1A | 3 × 3 × 3 | Mish | 256 × 256 × 1 |
| Convolution 1B | 3 × 3 × 3 | Mish | 256 × 256 × 32 |
| Max Pool | 2 × 2, stride 3 | 256 × 256 × 32 | |
| Convolution 2A | 3 × 3 × 3 | Mish | 128 × 128 × 32 |
| Convolution 2B | 3 × 3 × 3 | Mish | 128 × 128 × 80 |
| Max Pool | 2 × 2, stride 3 | 64 × 64 × 80 | |
| Convolution 3A | 3 × 3 × 3 | Mish | 64 × 64 × 160 |
| Convolution 3B | 3 × 3 × 3 | Mish | 64 × 64 × 160 |
| Max Pool | 2 × 2, stride 3 | 32 × 32 × 160 | |
| Bi-direction | 2D × 5 | ReLu | 1.25 × 105 |
| Convolution 4A | 3 × 3 × 3 | ReLu | 32 × 32 × 320 |
| Convolution 4B | 3 × 3 × 3 | ReLu | 32 × 32 × 320 |
| Up Convolution 4B | 2 × 2 | 64 × 64 × 320 | |
| Concat | Conv4b, Conv3b | 64 × 64 × 480 | |
| Convolution 5A | 3 × 3 × 3 | ReLu | 64 × 64 × 160 |
| Conv5B | 3 × 3 × 3 | ReLu | 64 × 64 × 160 |
| Up Convolution 5B | 2 × 2 | 128 × 128 × 160 | |
| Concat | Conv5b,Conv2b | 128 × 128 × 240 | |
| Convolution 6A | 3 × 3 × 3 | Mish | 128 × 128 × 80 |
| Convolution 6B | 3 × 3 × 3 | Mish | 128 × 128 × 80 |
| Up Convolution 6B | 2 × 2 | 256 × 256 × 80 | |
| Concat | Conv6B, Conv1B | 256 × 256 × 112 | |
| Convolution 6A | 3 × 3 × 3 | Mish | 256 × 256 × 32 |
| Convolution 6B | 3 × 3 × 3 | Mish | 256 × 256 × 32 |
| Convolution 7 | 3 × 3 × 3 | 256 × 256 × 2 |
Fig. 5Training accuracy vs. validation accuracy on LUNA16 trail set
Fig. 6Flowchart of the proposed DB-NET model
Fig. 7Histogram of the LUNA16 dataset and nodule sizes
Illustration of the various lung nodules present in the LUNA16 dataset
| S. no | Nodule type | Nodule image |
|---|---|---|
| 1 | Small node |
|
| 2 | GGO node |
|
| 3 | Calcific node |
|
| 4 | Cavitary |
|
| 5 | Juxta-vascular |
|
| 6 | Juxta-pleural |
|
| 7 | Isolated |
|
Ablation study on the LUNA16 testing set using the U-NET model
| S. no | Method | Dice coefficient (%) | Sensitivity (%) | Positive predictive value (%) |
|---|---|---|---|---|
| 1 | U-NET | 77.84 ± 21.79 | 78.98 ± 25.53 | 83.54 ± 22.55 |
| 2 | U-NET + BFPN | 81.22 ± 23.02 | 79.89 ± 25.85 | 84.89 ± 22.89 |
| 3 | U-NET + ReLU | 78.84 ± 12.52 | 84.45 ± 13.56 | 77.32 ± 14.45 |
| 4 | U-NET + ReLU + BFPN | 79.22 ± 12.36 | 91.69 ± 13.78 | 77.94 ± 14.35 |
| 5 | DB-NET | 88.89 ± 11.71 | 90.24 ± 13.15 | 77.92 ± 17.89 |
Fig. 8Lung CT scan ratio of each image on the LUNA16 dataset
Fig. 9The visual segmentation of the proposed algorithm on various heterogeneities of lung nodules
The segmentation results of the proposed architecture
| Performance | Benchmark LUNA16 dataset | |||
|---|---|---|---|---|
| When | When | Greater than 6 mm images | Less than 6 mm images | |
| Dice coefficient | 88.89 ± 11.71 | 90.24 ± 13.15 | 77.92 ± 17.89 | 75.63 ± 16.98 |
The measurable segmentation outcomes of the proposed model compared to other comparative models
| Authors | Architectures | Dice coefficient (%) | Sensitivity (%) | Positive predictive value (%) |
|---|---|---|---|---|
| Zhitao Xiao et al. (2020) | 3D-Res2U-NET | 81.22 ± 22.02 | 79.89 ± 24.85 | 84.89 ± 22.89 |
| Raghavendra Selvan et al. (2019) | U-NET-GNN | 78.84 ± 12.52 | 84.45 ± 13.56 | 77.32 ± 14.45 |
| Pius Kwao Gadosey et al. (2020) | Stripped-Down U-NET (SD-UNET) | 79.22 ± 12.36 | 91.69 ± 13.78 | 77.94 ± 14.35 |
| Sirojbek Safarov et al. (2021) | A-DenseUNet | 80.23 ± 23.02 | 77.88 ± 24.85 | 79.89 ± 22.89 |
| S Niranjan Kumar et al. (2021) | U-NET | 77.84 ± 21.79 | 78.98 ± 24.53 | 82.54 ± 21.55 |
| Kadia, Dhaval Dilip et al. (2021) | Advanced U-NET | 79.22 ± 22.02 | 79.89 ± 24.85 | 81.89 ± 22.89 |
| Dina M. Ibrahim et al. (2021) | ResNet152V2 + Gated Recurrent Unit (GRU) | 78.22 ± 22.02 | 79.89 ± 24.85 | 82.89 ± 22.89 |
| Proposed work | Proposed DB-NET Architecture | 88.89 ± 11.71 | 90.24 ± 13.15 | 77.92 ± 17.89 |
Features selected by PySckit Library concerning performance with the proposed architecture
| Model | Indices of selected features | Features | Dice coefficient | Mel frequency cepstral coefficient |
|---|---|---|---|---|
| 3D-Res2UNET | [3,10,13,28,29,32] | [12,13,14] | 81.22 ± 22.02 | 5 |
| UNET-GNN | [1,3,6,9,17,19,30] | [12,13,14] | 78.84 ± 12.52 | 7 |
| Stripped Down UNET (SD-UNET) | [2,5,9,5,8,9,16,19] | [12,13,14] | 79.22 ± 12.36 | 11 |
| A Dense U-NET | [1,3,5,8,23,27,28] | [12,13,14] | 80.23 ± 23.02 | 6 |
| U-NET | [3,10,13,28,29,32] | [12,13,14] | 77.84 ± 21.79 | 5 |
| Advanced U-NET | [3,10,13,28,29,32] | [12,13,14] | 79.22 ± 22.02 | 5 |
| ResNet15V2_Gated Recurrent Unit (GRU) | [1,3,6,9,17,19,30] | [12,13,14] | 78.22 ± 22.02 | 7 |
| Proposed DB-NET Architecture | [3,10,13,28,29,32] | [12,13,14] | 88.89 ± 11.71 | 6 |