| Literature DB >> 32906819 |
Satya P Singh1,2, Lipo Wang3, Sukrit Gupta4, Haveesh Goli4, Parasuraman Padmanabhan1,2, Balázs Gulyás1,2,5.
Abstract
The rapid advancements in machine learning, graphics processing technologies and the availability of medical imaging data have led to a rapid increase in the use of deep learning models in the medical domain. This was exacerbated by the rapid advancements in convolutional neural network (CNN) based architectures, which were adopted by the medical imaging community to assist clinicians in disease diagnosis. Since the grand success of AlexNet in 2012, CNNs have been increasingly used in medical image analysis to improve the efficiency of human clinicians. In recent years, three-dimensional (3D) CNNs have been employed for the analysis of medical images. In this paper, we trace the history of how the 3D CNN was developed from its machine learning roots, we provide a brief mathematical description of 3D CNN and provide the preprocessing steps required for medical images before feeding them to 3D CNNs. We review the significant research in the field of 3D medical imaging analysis using 3D CNNs (and its variants) in different medical areas such as classification, segmentation, detection and localization. We conclude by discussing the challenges associated with the use of 3D CNNs in the medical imaging domain (and the use of deep learning models in general) and possible future trends in the field.Entities:
Keywords: 3D convolutional neural networks; 3D medical images; classification; detection; localization; segmentation
Mesh:
Year: 2020 PMID: 32906819 PMCID: PMC7570704 DOI: 10.3390/s20185097
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Year-wise number of publications in PubMed while searching for ‘deep learning + medical’ and ‘3D deep learning + medical’ in the title and abstract in PubMed publication database (as at 1st July 2020).
Figure 2Criteria for literature selection for systematic review according to preferred reporting items for systematic seviews and meta-analyses (PRISMA) [23] guidelines.
Figure 3Typical architecture of 3D CNN.
Figure 4A typical architecture of AlexNet [14].
Figure 5(a) The intuition behind the inception (V1) module in GoogLeNet. Screening local clusters with 1 × 1 convolutional operations, screening spread-out clusters with 3 × 3, screening even more spread-out clusters with 5 × 5 convolutional operations, and finally conceiving the inception module by concatenating (b) residual building block in ResNet.
Figure 6The baseline architecture of 3D convolution neural network (CNN) for lesion segmentation. The figure is slightly modified from [54].
3D CNN for brain tumor/lesion segmentation on brain tumor segmentation (BRAST) challenges.
| Ref. | Methods | Data | Task | Performance Evaluation |
|---|---|---|---|---|
| Zhou et al. [ | A 3D variant of FusionNet (One-pass Multi-task Network (OM-Net)) | BRATS 2018 | brain tumor segmentation | 0.916 (WT), 0.827 (TC), 0.807(EC) |
| Chen et al. [ | Separable 3D U-Net | BRATS 2018 | --do-- | 0.893(WT), 0.830(TC), 0.742(EC) |
| Peng et al. [ | Multi-Scale 3D U-Nets | BRATS 2015 | --do-- | 0.850(WT), 0.720(TC), 0.610(EC) |
| Kayalıbay et al. [ | 3D U-Nets | BRATS 2015 | --do-- | 0.850 (WT), 0.872(TC), 0.610(EC) |
| Kamnitsas et al. [ | 11 layers deep 3D CNN | BRATS 2015 and ISLES 2015 | --do-- | 0.898 (WT), 0.750 (TC), 0.720(EC) |
| Kamnitsas et al. 2016 [ | 3D CNN in which features extracted by 2D CNNs | BRATS 2017 | --do-- | 0.918 (WT), 0.883(TC), 0.854 (EC) |
| Casamitjana et al. [ | 3D U-Net followed by fully connected 3D CRF | BRATS 2015 | --do-- | 0.917(WT), 0,836(TC), 0.768(EC) |
| Isensee et al. [ | 3D U-Nets | BRATS 2017 | --do-- | 0.850(WT), 0.740(TC), 0.640(EC) |
3D CNN for classification tasks in medical imaging.
| Ref. | Task | Model | Data | Performance Measures |
|---|---|---|---|---|
| Yang et al. [ | AD classification | 3D VggNet, 3D Resnet | MRI scans from ADNI dataset (47 AD, 56 NC) | 86.3% AUC using 3D VggNet and 85.4% AUC using 3D ResNet |
| Kruthika et al. [ | --do-- | 3D capsule network, 3D CNN | MRI scans from ADNI dataset (345 AD, NC, 605, and 991MCI) | Acc. for AD/MCI/NC 89.1% |
| Feng et al. [ | --do-- | 3D CNN + LSTM | PET + MRI scans from ADNI dataset (93 AD, 100 NC) | Acc. 65.5% (sMCI/NC), 86.4% (pMCI/NC), and 94.8 % (AD/NC) |
| Wegmayr et al. [ | --do-- | 3D CNN | ADNI and AIBL data sets, 20000 T1 scans | Acc. 72% (MCI/AD), 86 % (AD/NC), and 67 % (MCI/NC) |
| Oh et al. [ | --do-- | 3D CNN +transfer learning | MRI scans from the ADNI dataset (AD 198, NC 230, pMCI 166, and sMCI 101) at baseline. | 74% (pMCI/sMCI), 86% (AD/NC), 77% (pMCI/NC) |
| Parmar et al. [ | --do-- | 3D CNN | fMRI scans from ADNI dataset | Classification acc. 94.85 % (AD/NC) |
| Nie et al. [ | Brain tumor | 3D CNN with learning supervised features | Private, 69 patient (T1 MRI, fMRI, and DTI) | Classification acc. 89.85 % |
| Amidi et al. [ | Protein shape | 2-layer 3D CNN | 63,558 enzymes from PDB datasets | Classification acc. 78% |
| Zhou et al. [ | Breast cancer | Weakly supervised 3D CNN | Private, 1537 female patient | Classification acc. 78% 83.7% |
Figure 7The basic procedure for lung nodule detection. The figure is modified from Reference [92].