| Literature DB >> 33919669 |
Xuejiao Pang1, Zijian Zhao1, Ying Weng2.
Abstract
At present, the application of artificial intelligence (AI) based on deep learning in the medical field has become more extensive and suitable for clinical practice compared with traditional machine learning. The application of traditional machine learning approaches to clinical practice is very challenging because medical data are usually uncharacteristic. However, deep learning methods with self-learning abilities can effectively make use of excellent computing abilities to learn intricate and abstract features. Thus, they are promising for the classification and detection of lesions through gastrointestinal endoscopy using a computer-aided diagnosis (CAD) system based on deep learning. This study aimed to address the research development of a CAD system based on deep learning in order to assist doctors in classifying and detecting lesions in the stomach, intestines, and esophagus. It also summarized the limitations of the current methods and finally presented a prospect for future research.Entities:
Keywords: artificial intelligence; computer-aided diagnosis system; deep learning; esophageal lesion; gastric lesion; gastrointestinal endoscopy; intestinal lesion
Year: 2021 PMID: 33919669 PMCID: PMC8069844 DOI: 10.3390/diagnostics11040694
Source DB: PubMed Journal: Diagnostics (Basel) ISSN: 2075-4418
Figure 1Architecture of a representative convolutional neural network (CNN).
Figure 2Examples of endoscopic images of normal esophagus, early esophageal cancer (EC), gastric cancer, and duodenal ulcer: (a) normal esophagus, (b) early EC, (c) gastric cancer, and (d) duodenal ulcer.
Figure 3Architecture of the method proposed by Cao et al. [5].
Figure 4Architecture of the Gastric Precancerous Disease Network (GPD Net) proposed by Zhang et al. [8].
Characteristics of models for the diagnosis of gastric cancer and Helicobacter pylori (HP) infection.
| Study | Aim | Method | Performance | Train Dataset | Test Dataset |
|---|---|---|---|---|---|
| Ikenoyama et al. (2021) [ | Comparison between CNN and endoscopists | CNN based on SSD | CNN/Endoscopist: | 10,474 early-stage gastric cancer images and 3110 advanced-stage gastric cancer images | 209 gastric cancer images and 2731 normal images |
| Hirasawa et al. (2018) [ | Detection | CNN based on SSD | Sensitivity: 92.2% | 13,584 gastric cancer images | 2296 gastric cancer images |
| Sakai et al. (2018) [ | Detection | CNN based on GoogLeNet | Accuracy: 87.6% | 9587 gastric cancer images and 9800 normal images | 4653 gastric cancer images and 4997 normal images |
| Cao et al. (2019) [ | Detection + segmentation | Mask R-CNN | AP: 61.2% | 1000 positive samples and 250 negative samples | 120 positive samples and 29 negative samples |
| Li et al. (2020) [ | Classification | CNN + M-NBI | Accuracy: 90.91% | 1702 gastric cancer images and 386 normal images | 170 gastric cancer images and 171 normal images |
| Shibata et al. (2020) [ | Detection | Mask R-CNN | Average Dice: 71.0% | 533 gastric cancer images and 1208 normal images | Five-fold cross-validation |
| Zhang et al. (2017) [ | Classification | GPD Net | Accuracy: 88.9% | 921 images of erosion, 918 images of polyps, and 944 images of ulcer | 300 images of erosion, 300 images of polyps, and 300 images of ulcer |
| Shichijo et al. (2017) [ | Classification | First CNN based on GoogLeNet | First/second | 32,208 images either positive or negative for HP | 11,481 images |
| Itoh et al. (2018) [ | Detection | CNN based on GoogLeNet | AUC: 95.6% | 596 images | 30 images |
| Nakashima et al. (2018) [ | Classification | CNN based on GoogLeNet | AUCs: 66.0% (WLI), | 648 images for each WLI, BLI-bright, and LCI | 60 separate images for WLI, BLI-bright, and LCI |
Figure 5Flowchart of the online and offline 3D model proposed by Yu et al. [22].
Figure 6Main framework of Y-Net proposed by Mohammed et al. [23].
Characteristics of models for the classification and detection of colon polyps.
| Study | Aim | Method | Performance | Train Dataset | Test Dataset |
|---|---|---|---|---|---|
| Tajbakhsh et al. (2015) [ | Detection | Three-way image presentation + CNN | Sensitivity: about 75.0% | 20 collected short colonoscopy videos (10 positive, 10 negative) | 20 collected colonoscopy videos (10 positive and 10 negative) |
| Yu et al. (2017) [ | Detection | Offline and Online 3D FCN | F1-score: 78.6% | ASU-Mayo clinic database (20 colonoscopy videos) | ASU-Mayo clinic database (18 short colonoscopy videos) |
| Mohammed et al. (2018) [ | Detection | Y-Net | F1-score: 85.9% | ASU-Mayo clinic database (20 colonoscopy videos) | ASU-Mayo clinic database (18 short colonoscopy videos) |
| Haj-Manouchehri et al. (2020) [ | Detection + segmentation | CNN based on VGG; | Detection: (accuracy) 86.0% | Two collected colonoscopy videos for detection; | 1 collected colonoscopy video for detection; CVC-CLINIC and ETIS-LARIB datasets for segmentation |
| Zhang et al. (2017) [ | Detection + classification | CNN based on CaffNet | Precision: 87.3% | Source: ImageNet database and Places205; | 50 images for each class (nonpolyp, hyperplasia, and adenoma polyps) + 10 images for each class (hyperplasia, serrated adenoma, and adenoma) for 5 trials |
Figure 7Framework of the method proposed by Fonollà et al. [38].
Characteristics of models for the diagnosis of esophageal squamous cell carcinoma (ESCC) and esophageal adenocarcinoma (EAC) in the esophagus.
| Study | Aim | Method | Performance | Train Dataset | Test Dataset |
|---|---|---|---|---|---|
| Horie et al. (2019) [ | Detection of ESCC + EAC | CNN based on SSD | Sensitivity: 97.0% (ESCC) | 8428 EC images | 1118 images (EC + normal) |
| Cai et al. (2019) [ | Detection of ESCC | DNN-CAD | Accuracy: 91.4% | 2428 (1332 abnormal and 1096 normal) esophagoscopic images | 187 images |
| Guo et al. (2020) [ | Detection of ESCC | CNN based on SegNet | AUC: 98.9% | 2770 images (precancerous and early-stage ESCC); 3703 images (noncancerous) | Dataset A: 1480 images (precancerous + ESCC); B: 5191 images (noncancerous); C: 27 videos (precancerous + ESCC); D: 33 videos (noncancerous) |
| Ohmori et al. (2020) [ | Detection of ESCC | CNN based on SSD | Performance is good; 100.0% sensitivity by non-ME and 98.0% by ME | 9591 non-ME + 7844 ME images from superficial ESCCs; 564 non-ME + 2744 ME images of noncancerous lesions; 1128 non-ME + 691 ME images of normal esophagus | 255 non-ME WLI images; 268 non-ME NBI/BLI images; 204 ME NBI/BLI images |
| Tokai et al. (2020) [ | Invasion depth of ESCC | CNN based on SSD (detection), GoogLeNet (estimation) | Detection: 95.5% | 1751 images of ESCC | 291 test images |
| Mendel et al. (2017) [ | Diagnosis of EAC | CNN based on ResNet | Sensitivity: 94.0% | 4157 noncancerous region patches and 3666 cancerous region patches | Leave-one-patient-out cross-validation |
| Hashimoto et al. (2020) [ | Detection of EAC | Model based on Xception and YOLO v2 | Accuracy: 95.4% | 916 images of BE (high-grade dysplasia/T1 cancer) and | 458 test images (225 dysplasia and 233 non-dysplasia) |
| Fonollà et al. (2019) [ | Diagnosis of EAC | Assemble of 3 DCNN based on VGG16 | AUC: 96.0% | 134 NDBE, 38 HGD/EAC regions; total 8772 images | 99 NDBE and 42 HGD/EAC; total of 7191 images |
| Ghatwary et al. (2019) [ | Comparison | CNN based on R-CNN/Fast R-CNN/Faster R-CNN/SSD | SSD is the best: | 50 images (EAC) and 50 images (noncancerous) before data augmentation | Leave-one-patient-out cross-validation |