| Literature DB >> 35814554 |
H Alaskar1, A Hussain2,3, B Almaslukh1, T Vaiyapuri1, Z Sbai1, Arun Kumar Dubey4.
Abstract
Recent revolutionary advances in deep learning (DL) have fueled several breakthrough achievements in various complicated computer vision tasks. The remarkable successes and achievements started in 2012 when deep learning neural networks (DNNs) outperformed the shallow machine learning models on a number of significant benchmarks. Significant advances were made in computer vision by conducting very complex image interpretation tasks with outstanding accuracy. These achievements have shown great promise in a wide variety of fields, especially in medical image analysis by creating opportunities to diagnose and treat diseases earlier. In recent years, the application of the DNN for object localization has gained the attention of researchers due to its success over conventional methods, especially in object localization. As this has become a very broad and rapidly growing field, this study presents a short review of DNN implementation for medical images and validates its efficacy on benchmarks. This study presents the first review that focuses on object localization using the DNN in medical images. The key aim of this study was to summarize the recent studies based on the DNN for medical image localization and to highlight the research gaps that can provide worthwhile ideas to shape future research related to object localization tasks. It starts with an overview on the importance of medical image analysis and existing technology in this space. The discussion then proceeds to the dominant DNN utilized in the current literature. Finally, we conclude by discussing the challenges associated with the application of the DNN for medical image localization which can drive further studies in identifying potential future developments in the relevant field of study.Entities:
Mesh:
Year: 2022 PMID: 35814554 PMCID: PMC9259335 DOI: 10.1155/2022/6347307
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1General architecture of the FCN.
Figure 2General architecture of U-Net.
Figure 3General architecture of LSTM.
Figure 4Trending chart depicting the temporal distribution of the chosen articles.
State-of-the-art literature on the DNN for localization of abnormalities.
| Reference | Adopted DNN | Dataset used | Research findings |
|---|---|---|---|
| Onthoni et al. [ | Pretrained CNN | 10,078 CT images from autosomal dominant polycystic kidney disease | mAP = 94% localization of polycystic kidney |
| Roy et al. [ | STN | Italian COVID-19 lung ultrasound database (ICLUS-DB), containing 277 LUS videos from 35 patients | F1-score of 65.1 for localizing COVID-19 lesion on LUS images |
| X. Wang et al. [ | U-Net | 630 CT scans | Accuracy of 0.901 for localizing of the COVID-19 lesions |
| Noothout et al. [ | FCN | Two datasets, one containing 61 olfactory MR scans and another consisting of 400 cephalometric X-rays from ISBI 2015 grand challenge images | Distance errors were 1.45 mm for the the right and 1.55 mm for the lift in the localization of the coronary ostium |
| Cano-Espinosa et al. [ | U-Net | CT images | DICE coefficients of 0.875, 0.914 for PMA and SFA, respectively, in CT images |
| W. Jiang et al. [ | ResU-Net | Dataset containing 60 cases of MR images from 30 patients available for MICCAI WMH segmentation challenge | DICE coefficient of 0.832 for localizing WMH |
| Zhou et al. [ | CNN | MRI images taken from 1537 patients | Dice distance of 0.501 ± 0.274 for breast cancer localization |
| Lin et al. [ | U-Net | MRI (169 patients with cervical cancer stage IB–IVA captured; diffusion-weighted (DW) images) | Positive predicted value 0.92 for cervical tumors localization in (MRI) |
| Trebeschi et al. [ | CNN | 140 patients with biopsy proven locally advanced rectal carcinoma (LARC) | Dice Coefficient 0.68 for localization of rectal cancer in MRI |
| Liang et al. [ | CNN | FFDM images for 779 positive cases and 3018 negative cases | Positive 0.52 localization of breast cancer |
| Naseer Bajwa et al. [ | CNN | Online retinal fundus image database for glaucoma analysis and research. High-resolution fundus (HRF) image database optical coherence tomography (OCT) and color fundus images of both the eyes of 50 healthy persons | Accuracy 100% for localized optic disc |
| Reena and Ameer [ | AlexNet | Microscopic blood images (257 cells belonging to five types of leukocytes) | Mean average precision 98.42% for leukocytes localization |
| Fan et al. [ | CNN | Four datasets, namely, BCSI, LISC, and two other medical datasets | Dataset 1 0.99544, Dataset 2 0.99432, BCSI 0.98947, and LISC 0.98443 |
| Huang et al. [ | VGG-16 | Diffusion kurtosis images of 59 patients with epilepsy lesions in the hippocampus | Accuracy 90% for localization of epileptic foci |
| Heutink et al. [ | CNNs | 123 temporal bone CT scans were acquired with two UHR-CT scanners | Error 8% for localization |
| Cheng et al. [ | CNN | PXR dataset (frontal pelvic radiograph) | Accuracy of 95.9% for localization of hip fractures on plain frontal pelvic radiographs |
| González-Gonzalo et al. [ | VGG-16 | The kaggle DR dataset with 35,126 images from 17,563 patients | False positive 0.71 for localization of diabetic retinopathy (DR) and age-related macular degeneration abnormalities |
| Mwikirize et al. [ | CNN | 2D B-mode US images | Localization of needles inserted both in-plane and out-of-plane US image |
| Man et al. [ | U-Net | NIH dataset with 82 contrast-enhanced abdominal CT images | Recall rate 0.9 for pancreas localization on CT images |
| Y. Q. Jiang et al. [ | GoogleNet | 8046 microscopic ocular images | Mean intersection over union 0.863 for localizing basal cell carcinoma |
| Roggen et al. [ | Mask R-CNN | X-ray images from 12 abdominal cancer patients | |
| Shen et al. [ | CNN | NYU breast cancer screening dataset | Accuracy 78.1% for localized malignant lesions |
| Joel et al. [ | CNN | 637 cone beam CT volumes | The mean curve distance 0.56 mm for localization of the mandibular canals |
| Winkler et al [ | CNN | Six dermoscopic image sets. Each set included 30 melanomas and 100 benign lesions | Accuracy 93.3% for melanoma localization |
| H. Wang et al. 2020 [ | CheXLocNet | SIIM-ACR pneumothorax segmentation dataset 2079 radiographs | Dice score of 0.72 for localized pneumothorax lesions in chest radiographs |
| Poon et al. [ | DL | 291,090 colonoscopy videos | Polyp-based sensitivity=96.9 % |
| Urban et al. [ | CNN | 8,641 images from screening colonoscopies collected from 2000 patients | Accuracy, 95% for polyp-localization |
| Ouyang et al. | DL | CXR (NIH ChestX-ray14 and CheXpert) | |
| Guan et al. [ | CNNs | ChestX-ray14 collects 112,120 frontal-view images of 30,805 patients [ | AUC 0.871 for localizing pneumonia infection in CXR |
| Rajaraman et al. [ | CNNs | Radiological Society of North America (RSNA) CXR dataset [ | mAP 0.317 for localizing abnormalities on CXR |
| Rajaraman [ | VGG-16 | CXR [ | Accuracy 93.6% for localizing pneumonia infection in CXRs |
| Kermani et al. [ | NF-R-CNN | 3250 axial CMR images for 65 patients with ARVD | Mean error 7.33 ± 8.1 for heart localization in cardiac MR images |
| Vaiyapuri et al. [ | DL | The dataset holds a total of 500 CT images, with 250 images of pancreatic tumor and 250 images of nonpancreatic tumor | Near-optimal ACC of 0.9840, and a max ACC of 0.9935 on CT images towards pancreatic tumor localization |
| Groves et al. [ | CNN | 3825 US images | RMSE of 0.62 and 0.74 mm in the axial and lateral, respectively |
| Xue et al. [ | CNN | IHC images of colon tissue | Accuracy, 92.69% for protein subcellular localization |
| Al Arif et al. [ | FCN | 296 lateral cervical spine X-ray images | Dice similarity coefficient of 0.94 for spine localization |
| Won et al. [ | CNN | MR images | Accuracy 77.5% for localizing the center position of the spine canal |
| Peña-Solórzano et al. [ | CNN | 3D whole body CT scans | Dice scores of 0.99, 0.96, and 0.98 were obtained in the axial, coronal, and sagittal views for femur localization |
| Goyal et al. [ | CNN | 1775 images of DFU | Mean average precision of 91.8% for diabetic foot ulcers localization (DFU) |
| Afshari et al [ | CNN | 479 imaging captures the metabolic activity of tissue (PET scans) taken from 156 patients [ | Localization error 14 mm for localized anatomical objects in PET scans |
| Sarikaya et al [ | CNN | Video data of ten surgeons performing six different surgical tasks | Precision of 91% for localization in robot-assisted surgery RAS videos |
| Davidson et al. [ | CNN | 290 images of 142 healthy retinas and 148 retinas afflicted by Stargardt disease, acquired from 8 subjects with Stargardt disease | Dice score of 0.9577 for cone localization in images of healthy retinas |
| Dolz et al. [ | U-Net | IVD dataset is 16 3D multimodal MRI images | Localization error was 0.4 |
| Arik et al. [ | CNN | Cephalometric X-ray image dataset which includes 19 anatomical landmarks | 75.58%, 75.37%, and 67.68% accuracy for localizing sella, gonion and articulate landmarks |
| van der putten et al. [ | CNN | 494,355 endoscopic images | Accuracy 92% |
Temporal distribution of primary articles accepted for the study.
| Year | Number of articles |
|---|---|
| 2020 | 29 |
| 2019 | 19 |
| 2018 | 16 |
| 2017 | 10 |
| 2016 | 12 |
| 2015 | 8 |
State-of-the-art literature on the DNN for localization of anatomical structure.
| Reference | Adopted DNN | Dataset used | Research findings |
|---|---|---|---|
| Baumgartner et al. [ | CNN | US video frame data with 1003 midpregnancy scans | Precision, 69% and recall, 80%, |
| Ghesu et al. [ | Adaptive DNN with marginal space learning | 3D US images collected from 869 patients with 2891 volumes | Accuracy, 45.2% |
| Chen et al. [ | CNN with transfer learning | 11,942 foetal abdominal US images collected from pregnant women | Accuracy, 89.6% |
| Chen et al. [ | Extended (Chen et al., 2015b) with LSTM to extract spatial temporal features | 11,942 foetal abdominal US images collected from pregnant women | Accuracy, 90.8% |
| Baumgartner et al. [ | CNN based on a VGG-16 model | 2694 2D US images taken from gestational women | Accuracy, 77.8% |
State-of-the-art literature on application of the DNN for landmark localization.
| Reference | Adopted DNN | Dataset used | Research findings |
|---|---|---|---|
| Yang et al. [ | CNN with shape statistics | Three sets of 2D MRI slices | Minimum error rate of 1.61% for landmark localization |
| Kong et al. [ | LSTM with a CNN with new loss function | MRI cardiac sequences collected from 420 patients | Average frame difference (aFD), 38% |
| Emad et al. [ | CNN with different kernel sizes | MRI cardiac sequences collected from 33 patients | Accuracy, 98.66% |
| de vos et al. [ | Fusion model with three CNNs | Dataset containing 100 low-dose CT images | Median dice score of 89% for heart |
| Yan et al. [ | CNN with multistage and multi-instances learning | Two datasets: first with synthetic dataset and later with whole body CT scan dataset | Accuracy, 89.8% |
| Lu et al. [ | Different DL architectures considering the orthogonal orientations | 499 patient CT body scans | Error rate < 2.0 on average for organ localization |
| de vos et al. [ | CNN with spatial pyramid pooling | Three different datasets with 200 chest CT, 100 cardiac CT angiography (CTA), and 100 abdomen CT scans, respectively | F1-score > 95% for all three datasets |
| Feng et al. [ | CNN extracts features at different scales for voxel-level nodule segmentation | Public LIDC-IDRI dataset | Accuracy, 88.4% |
| Wolterink et al. [ | Leverages CNN for feature extraction and Random Forest for coronary artery calcification | Cardiac CT angiography (CCTA) from 50 patients | Accuracy, 0.8 |
| Hwang and kim [ | STL with joint optimization of classification and localization simultaneously | Two datasets, namely, CXR and mammogram | Accuracy, 83.69 |