| Literature DB >> 35079207 |
Rammah Yousef1, Gaurav Gupta1, Nabhan Yousef2, Manju Khari3.
Abstract
Medical images are a rich source of invaluable necessary information used by clinicians. Recent technologies have introduced many advancements for exploiting the most of this information and use it to generate better analysis. Deep learning (DL) techniques have been empowered in medical images analysis using computer-assisted imaging contexts and presenting a lot of solutions and improvements while analyzing these images by radiologists and other specialists. In this paper, we present a survey of DL techniques used for variety of tasks along with the different medical image's modalities to provide critical review of the recent developments in this direction. We have organized our paper to provide significant contribution of deep leaning traits and learn its concepts, which is in turn helpful for non-expert in medical society. Then, we present several applications of deep learning (e.g., segmentation, classification, detection, etc.) which are commonly used for clinical purposes for different anatomical site, and we also present the main key terms for DL attributes like basic architecture, data augmentation, transfer learning, and feature selection methods. Medical images as inputs to deep learning architectures will be the mainstream in the coming years, and novel DL techniques are predicted to be the core of medical images analysis. We conclude our paper by addressing some research challenges and the suggested solutions for them found in literature, and also future promises and directions for further developments.Entities:
Keywords: Deep learning (DL); Medical data augmentation; Medical imaging; Transfer learning
Year: 2022 PMID: 35079207 PMCID: PMC8776556 DOI: 10.1007/s00530-021-00884-5
Source DB: PubMed Journal: Multimed Syst ISSN: 0942-4962 Impact factor: 2.603
Fig. 1Deep learning implementation and traits for medical imaging application
Fig. 2DL basic categories as per paper organization
Fig. 3Basic common deep learning architectures. A Restricted Boltzmann machine. B Recurrent Neural Network (RNN). C Autoencoders. D GANs
Fig. 4The basic models used in medical imaging: A ResNet architecture, B U-Net architecture [75], C CNN AlexNet architecture for breast cancer [76], and D Dense Net architecture [77]
Fig. 5Timeline of mostly used DL models in medical imaging
Fig. 6Surveyed DL applications in medical imaging
Fig. 7Liver tumor segmentation using CNN architecture [86]
Fig. 8The number of challenges related to segmentation in medical imaging from 2007 to 2020 listed on Grand Challenges regarding the imaging modalities
DL models for medical images’ segmentation
| References | Algorithm | Organ (substructure) | Dataset | Performance | Citation |
|---|---|---|---|---|---|
| Dong et al. [ | U-Net FCN | Brain (tumor) | BRATS-2015 | DSC = 0.88, 0.87, 0.81 For complete, core, and enhancing tumor, respectively | 427 |
| Soltaninejad et al. [ | Random Forest + FCN for feature learning | Brain (tumor) | BRATS-2013 | DSC = 0.88, 080 and 0.73 for complete tumor, core, and enhancing tumor, respectively | 14 |
| Havaei et al. [ | 2D-CNN (DNN) | Brain | BRATS-2015 | DSC = 0.88, 0.79, 0.73 For complete, core, and enhancing tumor, respectively | 1974 |
| Wang et al. [ | Cascaded CNN with residual connections | Brain (tumor) | BRATS-2017 | DSC = 0.87, 0.77, 0.78 For complete, core, and enhanced tumor, respectively | 296 |
| Chen et al. [ | DeconvNet | Brain (ischemic lesions) | Locally collected from 761 patient | Dice coefficient = 0.88 | 149 |
| Li et al. [ | GAN | Brain (tumor) | BRATS-2017 | DSC = 0.87, 0.72, 0.68 For complete, core, and enhanced tumor | 32 |
| Zhou et al. [ | 3D variant of FusionNet (One-pass Multi-task Network (OM-Net)) | Brain (tumor) | BraTS-2018 | DSC = 0.916 (WT), 0.827 (TC), 0.807 (EC) | 34 |
| Isensee et al. [ | 3d-Unets | Brain (tumor) | BraTS-2017 | DSC = 0.850 (WT), 0.740 (TC), 0.640 (EC) | 300 |
| Korfiatis et al. [ | AutoEncoder | Brain (white matter hyperintensities) | BRATS-2015 + locally collected glioma dataset | Dice coefficient = 0.88 | 48 |
| Liu et al. [ | landmark-based deep multi-instance learning (LDMIL) | Brain (Alzheimer) | 1526 MRIs (ADNI-1, ADNI-2, MIRIAD) | F-score = 0.33 Accuracy = 76.9% | 160 |
| Liskowski et al. [ | DNN | Eye | DRIVE, STARE, and CHASE databases | ROC = 0.99 Sensitivity > 0.87 | 593 |
| Fang et al. [ | CNN with graph search (CNN-GS) | Eye | 60 OCT volumes from 20 patients | – | 315 |
| Shankaranarayana et al. [ | Conditional GAN, ResU-net | Eye | RIM-ONE dataset | F-score (disc = 0.97, cup = 0.94) IOU (disc = 0.89, cup = 0.76) | 68 |
| Fu et al. [ | M-Net and Polar Transformation | Eye | ORIGA | AUC = 0.8508 SCES dataset: AUC = 0.8998 | 337 |
| Son et al. [ | U-Net, GAN | Eye | DRIVE, STARE Databases | DRIVE dataset: DSC = 0.829 ROC = 0.9803 STARE dataset, DSC = 0.834 ROC = 0.9838 | 117 |
| Li et al. [ | Custom CNN | Liver (tumor) | 30 CT scans | Dice coefficient = 0.80 | 150 |
| Peijun Hu et al. [ | Custom 3D-CNN | Liver (structure) | 121 CT scans locally collected + 30 scans from SILVER07 dataset | Average score = 80.3 ± 4.5 | 129 |
| Ben Cohen et al. [ | FCN | Liver (lesion) | 20 CT scans locally collected + 20 CT scans from SILVER07 Dataset | Average Dice index of 0 | 150 |
| Yang et al. [ | deep image-to-image network with Adversarial Training (DI2IN-AN) | Liver (structure) | 1000 3D-CT volumes | Mean Dice Score = 0.95 | 159 |
| Cheng et al. [ | Stacked Denoising Auto Encoder (SDAE) | Breast | 1400 CT scans locally collected | Accuracy = (82.4 ± 4.5) % | 528 |
| Al-Antari et al. [ | YOLO + full resolution convolutional network (FrCN) | Breast | 410 X-ray mammography images from (INbreast database) | Accuracy = 98.96%, MCC = 97.62%, F1-score = 99.24% AUC = 93.05% | 128 |
| Skourt et al. [ | U-Net | Lung (parenchyma (nodule) segmentation) | (LIDC-IDRI) dataset | dice coefficient index = 0.9502 | 115 |
| Kalinovsky et al. [ | Encoder-Decoder CNN(ED-CNN) | Lung | 354 X-Ray chest images | Min-Dice score = 0.926 Max-Dice score = 0.974 Accuracy = 0.96 | 46 |
| Roy et al. [ | Custom CNN (Reg-STN) (CNN + Reg-STN + SORD model) SORD: Soft ORDinal regression Regularized Spatial Transformer Networks (Reg-STN) | Lung (COVID-19) | (ICLUS-DB) 277 lung ultrasound (LUS) videos from 35 patients, | Accuracy = 96, Binary Dice score = 0.75 F1-score = 71.4% | 131 |
| Murphy et al. [ | U-Net and CNN | Lung (COVID-19) | Private CXR images (24 678 chest radiographs) | AUC = 0.81 | 101 |
| Kline et al. [ | Custom CNN | Kidney | 2400 MRI locally generated images | Dice score = 0.97 ± 0.01 Jaccard score = 0.94 ± 0.03 | 67 |
| Jinlian Ma et al. [ | Custom CNN | Thyroid (nodule) | US images from 250 patient | Dice coefficient = 0.92 | 63 |
| Zhang et al. [ | multiple supervised residual network (MSRN) | Bone (osteosarcoma) | 1900 CT images from 15 osteosarcoma patients | Dice similarity coefficient (DSC) = 89.22%, a sensitivity = 88.74% and a F1-measure = 0.9305 | 17 |
| Li Yu et al. [ | Custom CNN | Heart (Fetal Left Ventricle) | 51 patients US | Dice coefficient = 94.5% | 60 |
| Jafari et al. [ | Custom CNN | Skin (melanoma) | 126 clinical images | Accuracy = 98.7% | 60 |
Fig. 9Lesion detection algorithm flowchart [118]
Deep learning applications to medical imaging detection
| References | Organ | CNN architecture | Dataset | Evaluation metrics | Citations |
|---|---|---|---|---|---|
| Zhang et al. [ | Brain (landmarks) | FCN | 3D T1-weighted (MR) images of 700 subjects(350/350) | Error (mean _ SD) = 2.94 _ 1.58 mm 1200 brain landmarks were detected | 109 |
| Zhang et al. [ | Prostate | FCN | 3D computed tomography (CT) images of 73 subjects | Error (mean _ SD) = 3:34 _ 2:55 7 prostate landmarks were detected | 109 |
| Nakao et al. [ | Brain (cerebral aneurysms) | Custom CNN | Locally collected MRI dataset | Accuracy = 94.2% | 74 |
| Tsehay et al. [ | Prostate (cancer) | DCNN | 196 MRI locally collected dataset (train = 144, test = 52) | 0.86 = AUC | 30 |
| Korsuk Sirinukunwattana et al. [ | Abdomen (nucleus) colon cancer) | Spatially Constrained Convolutional Neural Network (SC-CNN) | 2000 + histology images | Recall = 0.827 F-score = 0.802 Median distance = 2.236 | 740 |
| Setio et al. [ | Lung (pulmonary nodule) | Custom CNN | 888 CT scans of LIDC-IDRI dataset | Competition Performance Metric score (CPM) = 0.827 AUC = 0.996 | 806 |
| Li et al. [ | Lung (COVID-19) | U-Net and ResNet-50 | 4356 chest CT exams from 3,322 patients | AUC = 0.96 | 695 |
| Eduardo Luz et al. [ | Lung (COVID-19) | EfficientNet | 13,569 X-ray images | overall accuracy of 93.9%, COVID-19 sensitivity of 96.8% | 46 |
| Kassania et al. [ | Lung (COVID-19) | DenseNet -121 and ResNet-50 | 50 Chest X-ray and 150 CT scans | DenseNet -121Accuracy = 99%, ResNet-50 accuracy = 98% | 67 |
| Wang et al. [ | Breast(cancer) | Autoencoder | Histopathology Camelyon16 dataset consists of a total of 400 whole slide images (WSIs) | AUC = 0.995 for classification score of 0.7051 for the tumor localization | 658 |
| Dou et al. [ | Lung (Pulmonary Nodule) | 3D-CNN | LUNA16 dataset 888 CT scans | CPM score = 0.827 | 357 |
| Rajpurkar et al. [ | Chest (Pneumonia) | CheXnet (CNN) | ChestX-ray14 dataset over 100,000 X-ray images | F1-Score = 0.435 (95% CI 0.387, 0.481), | 1184 |
| Jinlian Ma et al. [ | Thyroid(nodules) | Cascade CNN | 21,523 ultrasound images | AUC = 98.51% confidence interval (CI) = 95% | 78 |
| Baka et al. [ | Vertebral | U-Net | Ultrasound and X-ray (train = 25 test = 19) | Recall = 0.88 F-means = 0.90 Accuracy = 92% | 40 |
| Alex et al. [ | Brain (Lesion) | GANs | MRI BraTS dataset | whole tumor dice score of 0.69, sensitivity of 91% | 45 |
| Bogunović et al. [ | Eye (Fluid) | 112 Optical coherence tomography (OCT) | AUC = 1.0 Accuracy = 90% | 49 | |
| Varun Gulshan et al. [ | Eye (diabetic retinopathy) | Inception-V3 | 1748 images from Messidor-2 dataset | AUC = 0.99, Sensitivity = 87%, Specificity = 98.50% | 3437 |
| Varun Gulshan et al. [ | Eye (diabetic retinopathy) | Inception-V3 | 9963 from EyePACS-1 dataset | AUC = 0.991, Sensitivity = 90.3 Specificity = 98.10 | 3437 |
| Hoo Chang Shin et al. [ | Thoraco-abdominal lymph node (LN) | Custom CNN | 983 2D CT from 176 patients | AUC = 0.94 for abdomen AUC = 0.95 for mediastinum | 3188 |
| Nasrullah et al. [ | Lung (Nodules) | (Custom CNN) 3D CMixNet with faster R-CNN | 888 CT scan low-dosed lung (LUNA16 dataset) | x free response receiver-operating characteristic (FROC) = 94.21% | 55 |
Fig. 10Classification of brain tumor using general CNN architecture
Deep learning applications to medical imaging classification
| References | Anatomic Organ | DL architecture | Dataset | Results evaluation metrics | Citations |
|---|---|---|---|---|---|
| Anthimopoulos et al. [ | Lung | Custom CNN | 120 ILD CT scans (ILD CT scans 109 HRCT scans) + (Bern University Hospital, “Inselspital” and consists of 26 HRCT scans of ILD cases) | Accuracy = 0.8561 Favg = 0.8547 | 836 |
| Nibali et al. [ | Lung (Pulmonary nodule) _ | ResNet | LIDC/IDRI dataset 1010 CT scans | Accuracy = 89.9 AUROC = 0.9459 | 136 |
| Christodoulidis et al. [ | Lung (tissues) | Custom CNN | (ALOT), (DTD), (FMD), (KTB), KTH-TIPS-2b, (HUG) Databases | F1-score avg = 0.8817 | 198 |
| Wu et al. [ | Lung (COVID-19) | ResNet-50 | Chest CT images of 495 patients | (AUC) = 0.732, accuracy = 0.700, sensitivity = 0.730 and specificity = 0.615 in validation set | 76 |
| Farid et al. [ | Lung (COVID-19) | Naïve Bayes | CT scans from online access Kaggle benchmark dataset,2020) | Accuracy of 96.07% | 37 |
| Pereira et al. [ | Lung (COVID-19) | CNN | CXR, a database, named RYDLS-20 | F1-Score of 0.89 | 161 |
| Huynh et al. [ | Breast | Custom CNN | Mammography, Private data (219 lesions) | AUC = 0.86 | 340 |
| Sun et al. [ | Breast (Cancer) | Custom CNN | 1874 mammogram images | Accuracy = 0.8243 AUC = 0.8818 | 160 |
| Swati et al. [ | Brain (tumor) | Custom CNN + transfer learning | 3064 Contrast Enhanced (CE-MRI) | Accuracy = 94.82% | 90 |
| Sajjad et al. [ | Brain (tumor) | Custom CNN | 3064 MRI | Accuracy = 95.5% | 180 |
| Deepak et al. [ | Brain (tumor) | Custom CNN | 3064 brain MRI images from 233 patients (figshare dataset) | Accuracy = 98% | 150 |
| Afshar et al. [ | Brain (tumor) | Capsule networks (CapsNet) | 3064 MRI images of 233 patients | Accuracy = 86.56% | 202 |
| Gao et al. [ | Brain (AD) | 3D-CNN | 285 datasets of 3D are collected from Navy General Hospital (CT scans) | Accuracy = 87.6% | 189 |
| Bharati et al. [ | Lung | 5606 X-Ray images | Accuracy = 73%, Fβ(0.5)-score = 0.68, Recall = 0.63, Precision = 0.69 | 34 | |
| Nasrullah et al. [ | Lung (Nodules) | Customized mixed link network 3D CMixNet + GBM + Biomarkers | 1018 CT scans (LIDC-IDRI dataset) | Accuracy = 94.17 | 52 |
| Zhou et al. [ | Breast (Cancer) | Weakly supervised 3D-CNN | Privately collected from 1537 female patients | Accuracy = 83.7% Confidence Interval (CI) = 95% | 35 |
| Zhang et al. [ | Breast (tumor) | PGBM-RBM-SVM | 227 shear-wave elastography (SWE) | Accuracy = 93.4%, Sensitivity = 88.6, Specificity = 97.1 AUC = 0.947 | 142 |
| Yang et al. [ | Brain (AD) | 3D VggNet, 3D Resnet | MRI scans from ADNI dataset (47 AD, 56 NC) | 86.3% AUC using 3D VggNet and 85.4% AUC using 3D ResNet | 58 |
| Schwyzer et al. [ | Lung (cancer) | ANN | 3936 PET slices | Standard dose): AUC = 0.989, sensitivity = 95.9%,, specificity = 98.1% (Ultralow dose): AUC = 0.970, sensitivity = 91.5%, specificity = 94.2% | 44 |
| de Carvalho Filho et al. [ | Lung | Custom CNN | 50,580 (14,184 malignant and 36,396 benign) nodules (LIDC-IDRI) | ACC = 92.63%, sensitivity = 90.7%, Specificity = 93.47% ROC = 0.934 | 33 |
| Shen et al. [ | Lung (cancer) | Multi-crop CNN MC-CNN | LIDC-IDRI (CT scans) dataset, 1010 patients | Accuracy = 87.14% AUC = 0.93, Sensitivity = 0.77, Specificity = 0.93 | 322 |
| Ozturk et al. [ | Lung (COVID-19) | Custom CNN (DarkCovidNet) | 481 CXR images | Accuracy = 98.08% for binary classification, Accuracy = 87.02% for multi-class cases | 679 |
| Ucar et al. [ | Lung (COVID-19) | SqueezNet | 5949 Chest X-ray scans (COVIDx) | Accuracy = 98.3% F1-score = 0.983, Matthew Correlation Coefficient (MCC) = 0.974 | 205 |
| Wang et al. [ | Lung (COVID-19) | Custom CNN (COVID-Net) | 13,975 CXR images across 13,870 patients | Accuracy94.3%, sensitivity = 96.8% | 925 |
| Rezvantalab [ | Skin (cancer) | DenseNet 201, ResNet152 | 10,135 dermoscopy skin images Actinic Keratoses (AK), Basal cell carcinoma (BCC), Squamous cell carcinoma (SCC) | AUC (ResNet-152) = (94.40%) F1-score (DenseNet) = (89.01–85.13%), ROC AUC (98.79–98.16%). And for Basal cell carcinoma ROC AUC (99.30%) | 36 |
| Dorj et al. [ | Skin (cancer) | ECOC SVM + CNN + Transfer learning of AlexNet | 3753 images | Accuracy 95.1%, Sensitivity = 98.9% Specificity = 94.17 | 111 |
| Esteva et al. [ | Skin (cancer) | Google’s Inception v3 CNN | 129,450 clinical images | Accuracy = 93.33% | 6173 |
| Awais et al. [ | Eye (Diabetic Macular Edema (DME)) | CNN- VGG16 | OCT (Optical Coherence Tomography) | Accuracy = 87.5%, SENSITIVITY = 93.5% SPECificity = 81% | 37 |
| Ting et al. [ | Eye (diabetic retinopathy | A deep learning system (DLS) | 40 752 images | AUC range was 0.889–0.983 AUC was 0.936 for (dataset: | 796 |
| Frid-Adar et al. [ | Liver (tumor) | GAN | 2D CT scans 182 cases (53 cysts, 64 metastases and 65 hemangiomas) | Sensitivity = 85.7%, Specificity = 92.4% | 348 |
Deep learning applications to medical imaging for registration
| Ref | Anatomical structure | DL architecture | Dataset | Transformation | Citations |
|---|---|---|---|---|---|
| Miao et al. [ | Chest | Custom CNN regression | Total Knee Arthroplasty (TKA) Kinematics X-ray + Virtual Implant Planning System (VIPS) + X-ray Echo Fusion (XEF) | Rigid | 295 |
| De vos et al. [ | Brain | Deep learning network for deformable image registration (DIRNet) | Sunnybrook Cardiac Data (SCD) contains 45 cardiac cine MRI + handwritten digits from the MNIST database | Deformable | 267 |
| Sun et al. [ | Brain | 3D-CNN | MRI (RESECT dataset) | Deformable | 15 |
| Chen et al. [ | Biogeography-based Optimization (BBO) algorithm | Retrospective Image Registration Evaluation Project (RIRE) with 6 modalities (CT, PET, MR-PD, MR-T1, MR-T2, MRMP-RAGE) | Geometric rigid transformation | 31 | |
| Niethammer et al. [ | Brain | CNN | MRI (LPBA (40) dataset 2D 3D CUMC (12), MGH(10), and IBSR(18) datasets) | Deformable | 31 |
| Wu et al. [ | Brain | CNN | MRI, LONI, ADNI databases | Deformable | 212 |
| Kang et al. [ | Abdomen | CNN | CT low-dose Grand Challenge | Contourlet transform + wavelet transform | 486 |
| Zhang et al. [ | Head, Abdomen, Chest | CNN | Private CT scans | Deformable | 142 |
| Stankevièius et al. [ | Eye | DNN | 3153 retinal images | Rigid | 6 |
| Haskins et al. [ | Prostate cancer | CNN | 679 MR images from the National Institute of Health (NIH) | Deformable | 53 |
Deep learning applications to medical imaging for characterization
| References | Anatomical site/task | Network | Dataset | Citation |
|---|---|---|---|---|
| Breast (Cancer risk assessment) | Customized AlexNet | Mammogram images (Train = 14,000, test = 1850) | 152 | |
| Breast (Cancer risk assessment) | Fine-tuned a pre-trained VGG16Net | Mammograms 604 images | 42 | |
| Breast (Diagnosis) | Pre-trained VGGNet 19 | Mammograms (245), MRI(690), US(1125) | 188 | |
| Samala et al. [ | Breast (Diagnosis) | AlexNet | Mammograms (train = 1545, test = 909) images | 115 |
| Breast (Solitary cyst diagnosis) | VGGNet customized | 1600 Mammograms lesions | 93 | |
| Anthimopoulos et al. [ | Lung (Interstitial disease) | Custom VGG | 14,696 CT patches | 849 |
| Lung (Pulmonary nodule staging) | DFCNet | LIDC-IDRI database, RIDER, LungCT-Diagnosis, and | 92 | |
| Lung (Interstitial disease) | Custom CNN | CT patches ( train = 36,106, test = 1050) from ALOT and KTH-TIPS-2b databases | 200 | |
| González et al. [ | Lung and chest (prognosis of chronic obstructive pulmonary disease (COPD)) | Custom CNN | CT scans (train = 7983, test = 1000 COPDGene + 1,672 ECLIPSE) | 112 |
| Brain (Survival) | CNN-S with transfer learning | Multiparametric MRI (112 patients | 284 |
Fig. 11Flowchart of medical images data handling
Public available datasets used for medical imaging
| Dataset/references | Anatomical organ | Image modality | No. of images/patients |
|---|---|---|---|
| BRATS [ | Brain | MRI | – |
| ADNI [ | Brain | MRI | 819 patients |
| ChestX-ray14 [ | Lung | X-ray | 30,000 + patients |
| LIDC-IDRI [ | Lung | CT | 1018 patients |
| ILD [ | Lung | CT | 120 patients |
| TCIA | Multiple organs with cancer | CT, MRI, PET | 30.9 million images/37,568 patients |
| DRIVE [ | Eye | SLO | 400 patients |
| STARE [ | Eye | SLO | 400 images |
| ISIC 2018 | Skin | JPEG | 2600 images |
| SILVER07 | Liver | CT | – |
| COVIDx [ | Lung | CXR | 13,975 CXR images across 13,870 |
Fig. 12Features’ extraction and selection types used for dimensionality reduction
Fig. 13ROC and AUC graph