Literature DB >> 33791323

Current Evidence and Future Perspective of Accuracy of Artificial Intelligence Application for Early Gastric Cancer Diagnosis With Endoscopy: A Systematic and Meta-Analysis.

Jiang Kailin1, Jiang Xiaotao1, Pan Jinglin2, Wen Yi3, Huang Yuanchen1, Weng Senhui1, Lan Shaoyang3, Nie Kechao1, Zheng Zhihua1, Ji Shuling1, Liu Peng1, Li Peiwu3, Liu Fengbin3.   

Abstract

Background & Aims: Gastric cancer is the common malignancies from cancer worldwide. Endoscopy is currently the most effective method to detect early gastric cancer (EGC). However, endoscopy is not infallible and EGC can be missed during endoscopy. Artificial intelligence (AI)-assisted endoscopic diagnosis is a recent hot spot of research. We aimed to quantify the diagnostic value of AI-assisted endoscopy in diagnosing EGC. Method: The PubMed, MEDLINE, Embase and the Cochrane Library Databases were searched for articles on AI-assisted endoscopy application in EGC diagnosis. The pooled sensitivity, specificity, and area under the curve (AUC) were calculated, and the endoscopists' diagnostic value was evaluated for comparison. The subgroup was set according to endoscopy modality, and number of training images. A funnel plot was delineated to estimate the publication bias. Result: 16 studies were included in this study. We indicated that the application of AI in endoscopic detection of EGC achieved an AUC of 0.96 (95% CI, 0.94-0.97), a sensitivity of 86% (95% CI, 77-92%), and a specificity of 93% (95% CI, 89-96%). In AI-assisted EGC depth diagnosis, the AUC was 0.82(95% CI, 0.78-0.85), and the pooled sensitivity and specificity was 0.72(95% CI, 0.58-0.82) and 0.79(95% CI, 0.56-0.92). The funnel plot showed no publication bias.
Conclusion: The AI applications for EGC diagnosis seemed to be more accurate than the endoscopists. AI assisted EGC diagnosis was more accurate than experts. More prospective studies are needed to make AI-aided EGC diagnosis universal in clinical practice.
Copyright © 2021 Kailin, Xiaotao, Jinglin, Yi, Yuanchen, Senhui, Shaoyang, Kechao, Zhihua, Shuling, Peng, Peiwu and Fengbin.

Entities:  

Keywords:  artificial intelligence; deep learning; early gastric cancer; endoscopy; machine learning

Year:  2021        PMID: 33791323      PMCID: PMC8005567          DOI: 10.3389/fmed.2021.629080

Source DB:  PubMed          Journal:  Front Med (Lausanne)        ISSN: 2296-858X


Introduction

Gastric cancer is ranked as the third leading cause of death from cancer worldwide (1). Most gastric cancers are diagnosed at advanced stages because their symptoms and signs tend to be inconspicuous and non-specific, leading to an overall poor prognosis, whereas in the case of early detection, the 5–years survival rate can exceed 90% (2–4). Endoscopic examination is still considered the most effective method for EGC detection (5). However, early gastric cancer (EGC) is particularly difficult to identify since it usually exhibits a subtle elevation or depression with faint redness, which is likely recognized as normal mucosa or gastritis. In addition, the invasion depth within the gastric wall is also hard to predict. Ten studies involving 3,787 patients who received an upper gastrointestinal endoscopy examination revealed an 11.3% miss rate of upper gastrointestinal cancers up to 3 years before diagnosis (6). A meta-analysis involving 2,153 lesion images showed that the area under the receiver operating characteristic curve (AUC) for the diagnosis of EGC using white light imaging (WLI) endoscopy was only 0.48 (7). In the past decade, the application of artificial intelligence (AI) in medicine has attracted extensive attention. AI-assisted endoscopic diagnosis is a hot spot of research. AI refers to the capacity of a computer to execute a task associated with intelligent beings, such as the “learn” function that mimics the cognitive ability of human beings (8). AI subfields contain machine learning and deep learning (Figure 1). Machine learning, a term originally created by Arthur Samuel in 1959, is a field of computer science, whereby a system is able to develop the ability to “learn” from the input data without a certain program (9). Common machine-learning methods in classification model training comprise ensemble trees, decision trees, support vector machines, k-nearest neighbors, etc. (10).
Figure 1

Artificial intelligence methods in medical imaging. Artificial intelligence (AI) methods for a typical classification task were shown. Two classical methods comprise traditional machine learning (A) and deep learning (B). Conv, Convolutional layer; Pool, Pooling layer; FC, receiver operating characteristic curve; EGC,: Early gastric cancer.

Artificial intelligence methods in medical imaging. Artificial intelligence (AI) methods for a typical classification task were shown. Two classical methods comprise traditional machine learning (A) and deep learning (B). Conv, Convolutional layer; Pool, Pooling layer; FC, receiver operating characteristic curve; EGC,: Early gastric cancer. Deep learning, which was initially applied in the image processing field in 1998, refers to the application of layers in non-linear processing based on machine learning algorithms used for feature extraction and transformation (11). Neural networks, similar to the human brain, particularly mimic closely interconnected neurons to recognize patterns, extract features or “learn” things about the input data to predict a result (12). Different model training paradigms, such as scaled-conjugate gradient, Levenberg-Marquardt and Bayesian regularization, have been termed “neural networks” (13). Several computer aided detection (CAD) algorithms for automatic early gastric cancer detection have been recommended for images from standard endoscopes. The performance improvements of original image classification models mainly depend on visual features and large-scale datasets, which are difficult to implement in EGC detection models. Although the invasion depth in EGC is defined differently, visual characteristics such as textures, colors, shapes, and regions are similar. To date, the existing data on the diagnostic value of AI for EGC diagnosis are scattered. Jin et al. (14) reviewed the current studies on AI application for gastric cancer, while the definite diagnostic ability of AI application for EGC was still unclear. The aim of this study was to systematically summarize the recent available studies on the diagnostic accuracy of AI on EGC diagnosis to address the current status of this area and discuss future perspectives.

Methods

Search Strategy and Study Selection

Electronic databases (PubMed, Web of Science, EMBASE, and the Cochrane Library) were searched from initiation to November 2020 using presupposed search terms. The following medical subject terms and keywords were used: “endoscopy,” “Endoscopic Diagnosis,” “early gastric cancer,” “artificial intelligence,” “computer-assisted diagnosis,” “Deep learning,” and “Convolutional neural network.” The full texts of potentially appropriate studies were then reviewed after the screenings of citations and abstracts exported from the electronic databases. The search strategy was shown as follows: (1) (artificial intelligence [Title/Abstract]) OR (computer-assisted diagnosis [Title/Abstract]) OR (Deep learning [Title/Abstract]) OR (Convolutional neural network [Title/Abstract]) (2) (endoscopy [Title/Abstract]) OR (Endoscopic Diagnosis [Title/Abstract]) OR (early gastric ancer [Title/Abstract]) (3) (1) AND (2).

Study Eligibility Criteria

The eligible studies fulfilled the following criteria: (1) the study was a diagnosis test about AI application in endoscopy for EGC diagnosis. Diagnosis test included AI detection of EGC from other gastric disease or distinguishment of invasion depth; (2) the absolute numbers of true-positive, false-negative, true-negative, and false-positive observations for EGC diagnosis were reported directly or were able to be calculated; (3) the study provided clear information about the database and number of images; (4) the study clearly described the CAD or CNN algorithms and the process applied in the EGC diagnosis.

Data Extraction

Two reviewers (Jiang X. T., Wen Y.) independently extracted information, including the author, publication year, region, study type, endoscopy modality, algorithm gold standard and dataset, and used the quality assessment of diagnostic accuracy studies-2 instrument to assess the quality of the study (15). Divergence was resolved through discussion and the involvement of the third reviewer (Li P. W.).

Statistical Analysis

Stata, version 14.2 (StataCorp, College Station, TX) was used for all statistical analyses. Graphpad Prism 8.2.1 was used to delineate the histogram. The TP, FP, FN, and TN observations of each study were input, and the pooled sensitivity and specificity with the 95% confidence intervals (CIs) for EGC diagnosis with AI were thus calculated. The forest plot was delineated. The inconsistency index (I2) test was used to evaluate the heterogeneity between studies using sensitivity (16). A fixed-effects model would be used with a I2 value <50%. More than 50% of the I2 values indicated significant heterogeneity. Under this situation, a random-effects model would be applied, and subgroup analysis and influence analysis were performed. A summary receiver operating characteristic (ROC) curve was plotted (17). The area under the curve (AUC) was calculated to estimate the diagnostic accuracy. When the AUC reaches 1.0, it suggests an excellent performance diagnostic test, while if the AUC approaches 0.5, it suggests a poor performance test. Publication bias was evaluated by the Deeks test.

Result

Literature Search and Characteristic of Studies

A total of 3,714 studies were retrieved after the search. After removing duplicated studies and excluding improper studies, 17 studies were reserved in this systematic analysis. While Ling et al. (18) distinguished differentiated and undifferentiated type EGC with a sensitivity and specificity of 88.6 and 78.6%, thus was finally excluded in our meta-analysis. A total of 16 studies were finally included in the meta-analysis according to the PRISMA flowchart (Supplementary Figure 1). Three studies were from Korea, eight studies were from Japan, four studies were from China, and one was from Pakistan. Nine studies used white light endoscopy (WLE) images to establish a training dataset, five studies used narrow band imaging (NBI) images, and two used both WLE and NBI images. Four studies distinguished the invasion depth of EGC. Seven studies compared the diagnostic ability of AI with endoscopists. Two studies applied video to train the dataset. No prospective studies were carried out currently. The general algorithm methods were Visual Geometry Group-16 (VGG-16), ResNet-50, GoogLeNet, Single Shot MultiBox Detector (SSD), Inception neural network and Support vector machines (SVM) classifier. Yoon et al. applied two kinds of algorithm models in his study. The basic characteristics of the included studies and the risk of bias using the Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) tool are presented in Table 1 and Supplementary Figure 2.
Table 1

Basic characteristic of the included studies.

YearNationStudy typeEndoscopy for trainingImage typeFormatProcessing image sizeDLAlgorithmAffiliated toolsGold standardTraining databaseEndoscopist involvementReal-time
Yoon et al. (19)2019KoreaRetrospectiveWLEImageNot mentionedNot mentionedCNNVGG-16(20)Grad-CAMWHO classification of Tumors (21), Japanese classification (22)Gangnam Severance Hospital, Yonsei University College of Medicine, KoreaNoNo
Cho et al. (23)2019KoreaRetrospectiveWLEImageJPEG1,280 × 640 pixelsCNNInception-Resnet-v2SGDHistopathologyEndoscopically biopsied or EMR/ESD lesions from Chuncheon and Dongtan Sacred Heart Hospitals,KoreaYesNo
Sakai et al. (24)2018JapanRetrospectiveWLEimageNot mentioned224 × 224 pixelsCNNGoogLeNet(25)NoHistopathologyNot mentionedNoNo
Horiuchi et al. (26)2019JapanRetrospectiveME-NBIImageNot mentioned224 × 224 pixelsCNNGoogLeNetNoHistopathologyCancer Institute Hospital, Ariake, Koto-ku, JapanNoNo
Lan et al. (27)2019ChinaRetrospectiveME-NBIImageNot mentioned299 × 299 pixels to 512 × 512 pixelsCNNInception-v3Keras deep learning frameworkRevisited Vienna classification of gastrointestinal epithelial neoplasia(28)Four hospitals in four areas of Zhejiang provinceYesNo
Toshiaki et al. (29)2018JapanRetrospectiveWLE, Chromoendoscopy and NBIimageNot mentioned300 × 300 pixelsCNNSSD(30)NoJapanese classificationCancer Institute Hospital Ariake, Japan, Tokatsu Tsujinaka Hospital, Japan and Tomohiro Institute of Gastroenterology and Proctology, Japan, Lalaport Yokohama Clinic, JapanNoNo
Yan et al. (31)2019ChinaRetrospectiveWLEimageNot mentioned299 × 299 pixelsCNNResNet50(32)NoJapanese classificationEndoscopy Center of Zhongshan Hospital, ChinaYesNo
Kanesaka et al. (33)2017JapanRetrospectiveME-NBIimageNot mentioned40 × 40 pixelsCADSVM classifierNopathology-proven EGCs resected by ESDEthics Committee of the Osaka International Cancer InstituteNoNo
Wu et al. (34)2018ChinaRetrospectiveWLE, NBI, BLEvideoNot mentioned224 × 224 pixelsCNNVGG-16, ResNet-50NoHistopathologyRenmin Hospital of Wuhan University, ChinaYesYes
Miyaki et al. (35)2013JapanRetrospectivemagnifying endoscopeimageNot mentioned1,280 × 1024 pixelsCADSVM classifierNoHistopathologyHiroshima University HospitalNoNo
Ikenoyama et al. (36)2020JapanRetrospectiveWLEimageNot mentioned300 × 300 pixelsCNNSSDSGDHistopathologyCancer Institute Hospital Ariake, Tokatsu-Tsujinaka Hospital, Tada Tomohiro Institute of Gastroenterology and Proctology, Lalaport Yokohama Clinic, JapanYesNo
Ali et al. (37)2018PakistanRetrospectiveChromoendoscopyImageNot mentionedNot mentionedCADSVM classifierG2LCM descriptorsNot mentionedPublic data-set at the Portuguese Institute of OncologyNoNo
Bun-Joo et al. (38)2020KoreaRetrospectiveWLEImageJPEG480 × 480 pixelsCNNInception-ResNet-v2 and DenseNet- 161Class activation map (CAM)HistopathologyChuncheon Sacred Heart HospitalNoNo
Horiuchi et al. (39)2020JapanRetrospectiveME-NBIVideoNot mentioned224 × 224 pixelsCNNGoogLeNetSGDHistopathologyLesions initially treated with ESD at the CancerInstitute HospitalYesNo
Ueyama et al. (40)2020JapanRetrospectiveME-NBIImageNot mentioned224 x 224 pixelsCNNResNet50SGDJapanese ClassificationDepartment of Gastroenterology, Juntendo University School of MedicineNoNo
Zhang et al. (41)2020ChinaRetrospectiveWLEImageNot mentionedNot mentionedCNNResNet34DeepLabv3 structureHistopathologyGastric cases admitted to Peking University People's HospitalYesYes

WLE, White Light Endoscopy; NBI, Narrow Band Imaging; BLI, blue-laser imaging; WHO, World Health Organization; SVM, support vector machine; SSD, Single Shot MultiBox Detector; CNN, Convolutional Neural Network, CAD, Computer-aided diagnosis; Grad-CAM, gradient-weighted class activation mapping; VGG-16, Visual Geometry Group-16, SVM, Support vector machines, SGD, Stochastic gradient descent.

Basic characteristic of the included studies. WLE, White Light Endoscopy; NBI, Narrow Band Imaging; BLI, blue-laser imaging; WHO, World Health Organization; SVM, support vector machine; SSD, Single Shot MultiBox Detector; CNN, Convolutional Neural Network, CAD, Computer-aided diagnosis; Grad-CAM, gradient-weighted class activation mapping; VGG-16, Visual Geometry Group-16, SVM, Support vector machines, SGD, Stochastic gradient descent.

Diagnostic Performance of AI on EGC Diagnosis

A total of 170,8519 images were utilized for machine training. A total of 22,621 EGC images from the 16 studies were included in the meta-analysis of EGC diagnosis. The diagnostic ability of AI-assisted endoscopy in each study is shown in Supplementary Table 1. The AUC of the AI-assisted endoscopy diagnosis in EGC detection was 0.96 (95% CI, 0.94–0.97) with heterogeneity I2 value of 0.98, thus the random effect model was applied. The pooled sensitivity was 86% (95% CI, 77–92%), and the specificity was 93% (95% CI, 89–96%). While the AUC, sensitivity and specificity of AI-assisted depth distinction was 0.82 (95% CI, 0.78–0.85), 72% (95% CI, 58–82%), and 79% (95% CI, 56–92%). The forest plots of sensitivity, specificity of AI detection and depth distinction are shown in Figures 2, 3. ROC of detection and depth distinction are shown in Figure 4. Influence analysis showed that Bum-Joo Cho, Hiroya Ueyama, and Yusuke Horiuchi's study had the greatest impact on the results (Supplementary Figure 3). After rejecting them, the pooled AUC, sensitivity and specificity were 0.95 (95% CI, 0.93–0.97), 85% (95% CI, 78–90%), and 92% (95% CI, 90–94%), respectively, which still indicated an accurate diagnostic ability of AI-aided diagnosis of EGC. The funnel plot asymmetry with a p-value of 0.81 showed the absence of publication bias for the included studies (Supplementary Figure 4).
Figure 2

The forest plot of pooled sensitivity and specificity of AI detection on EGC. The pooled sensitivity was 86% (95% CI, 77–92%) and specificity was 93% (95% CI, 89–96%).

Figure 3

The forest plot of pooled sensitivity and specificity of AI distinction depth on EGC. The pooled sensitivity was 72% (95% CI, 58–82%) and specificity was 79% (95% CI, 56–92%).

Figure 4

Area under the receiver operating characteristic curve (A). The AUC of the AI-assisted endoscopy diagnose in the EGC detection was 0.96 (95% CI, 0.94–0.97). (B) The AUC of the AI-assisted endoscopy diagnose in the EGC depth distinction was 0.82 (95% CI, 0.78–0.85).

The forest plot of pooled sensitivity and specificity of AI detection on EGC. The pooled sensitivity was 86% (95% CI, 77–92%) and specificity was 93% (95% CI, 89–96%). The forest plot of pooled sensitivity and specificity of AI distinction depth on EGC. The pooled sensitivity was 72% (95% CI, 58–82%) and specificity was 79% (95% CI, 56–92%). Area under the receiver operating characteristic curve (A). The AUC of the AI-assisted endoscopy diagnose in the EGC detection was 0.96 (95% CI, 0.94–0.97). (B) The AUC of the AI-assisted endoscopy diagnose in the EGC depth distinction was 0.82 (95% CI, 0.78–0.85).

Other Factors That Have an Impact on the Accuracy of AI

The effects of the original images from WLE or NBI on the AI diagnostic ability were compared. The sensitivity of the NBI image application was 95% (95% CI, 91–97%), while that of WLE was 73% (95% CI, 57–85%), and the specificity was 96% (95% CI, 70–100%) and 93% (95% CI, 90–95%). When the number of training images was more than 10,000, the sensitivity and specificity were 88% (95% CI, 83–92%) and 94% (95% CI, 91–96%), respectively, more than that of the sensitivity 85% (95% CI, 69–93%) and specificity 93% (95% CI, 82–97%) of the group that had >10,000 training images. For the control group, sensitivity and specificity of the expert endoscopist vs. non-expert endoscopist diagnosis were 79% (95% CI, 61–90%) vs. 73% (95% CI, 61–82%), 85% (95% CI, 77–90%) vs. 83% (95% CI, 67–92%), respectively. Here, the general expert endoscopists were those who had clinical experience with endoscopy examination for more than 10 years. Figure 5 shows the subgroup results.
Figure 5

Result of subgroup analysis. (A) The pooled sensitivity and specificity of number of images in training process showed when the images were more than 10,000, the diagnostic value would be better. (B) The pooled sensitivity and specificity of AI detection, expert endoscopist, and non-expert endoscopist showed AI detection and expert endoscopist judgement were significantly more accurate than non-expert endoscopist. (C) The pooled sensitivity and specificity of original images extracted by NBI and WLE showed NBI image applied performed better.

Result of subgroup analysis. (A) The pooled sensitivity and specificity of number of images in training process showed when the images were more than 10,000, the diagnostic value would be better. (B) The pooled sensitivity and specificity of AI detection, expert endoscopist, and non-expert endoscopist showed AI detection and expert endoscopist judgement were significantly more accurate than non-expert endoscopist. (C) The pooled sensitivity and specificity of original images extracted by NBI and WLE showed NBI image applied performed better.

Discussion

Japanese researchers published a minimum required standard for the “systematic screening protocol for the stomach,” which comprised 22 images of the stomach to precisely discover suspicious cancerous lesions (42). In 2016, the European Society of Gastrointestinal Endoscopy (ESGE) published a protocol comprising 10 images of the stomach (43). However, these protocols could not be carried out absolutely, and endoscopists may miss some regions during the examination due to individual operative levels and subjective factors, which can lead to the misdiagnosis of EGC (44–46). Deep learning (47, 48), which is typically based on artificial neural networks, aims at learning multilevel manifestations of data to make predictions. The development of deep convolutional neural networks has particularly altered the computer vision field (49, 50). Application of AI recognition with endoscopic images to detect the depth of wall invasion of gastric cancer was initially reported by Keisuke Kubota with an accuracy of 64.7% (51). Soon afterwards, several studies have shown excellent results for advanced technology. Hence, it is necessary to summarize the existing studies to realize the probable ability of AI on EGC detection and discuss what factors may influence the results. This is the first meta-analysis on the performance of AI on EGC diagnosis with endoscopy. In this article, we indicated that the application of AI in endoscopic detection of EGC achieved an AUC of 0.96 (95% CI, 0.94–0.97), a sensitivity of 86% (95% CI, 77–92%), and a specificity of 93% (95% CI, 89–96%), which manifested a more accurate diagnostic ability than independent detection by endoscopists, while the depth distinction was dissatisfied with a sensitivity, specificity and AUC of 0.82 (95% CI, 78–85%), 72% (95% CI, 58–82%), and 79% (95% CI, 56–92%). The common reasons for misdiagnosis were lesions of gastritis or flat or depressed texture and anatomical structure which was hard to identify. The cancer invasion depth was classically distinguished by morphologically evaluating several findings such as the concentration of stomach wall folds, the marginal ridge, the elasticity and thickness of the lesion, and the presence of variant of the stomach wall due to the volume of insufflation air in the stomach with WLE (52–54). Furthermore, the accuracy of discriminating EGC depth by conventional endoscopy was reported to be 62–80% (55). Thus, the AI applied endoscopy performed well on EGC depth determination. Bum-Joo Cho, Hiroya Ueyama and Yusuke Horiuchi's study (23, 26, 40) showed significant heterogeneity. Cho et al. used the Inception-Resnet-v2 model with an AUC of 74.5 (95% CI, 67.9–80.4) and a sensitivity of 28.3 (95% CI, 16.0–43.5). The included poor-quality images, composition of the database, and pathological classification criteria may cause poor diagnostic performance. In addition, we performed several subgroup analyses to delineate the probable influencing factors of AI performance. For the algorithm model, Simonyan et al. (56) investigated the value of the convolutional network depth on its accuracy in large-scale image recognition setting. The result showed that when the depth was pushed to 16–19 weight layers, it would have a significant improvement on the prior-art configurations. VGG-16 had 16 convolutional and three fully connected layers, which were carried out by five max-pooling layers and used filters with a small receptive field to achieve a low error rate in practice. On the other hand, SVM also performed excellently in the included studies. SVM is utilized in distinguishing two classes and creating the boundary line to maximize the distance between the hyperplane and the nearest sample. Compared to other mathematical models (57–59), SVMs are utilized to model physical systems by adapting their parameters (60–63). SVMs are widely known for their application in classification (64). The endoscopic image modality of validation set should be same to the training set. For training images from different endoscopy modalities, the sensitivity of studies using images from NBI seemed to be better than those using images from WLE (96 vs. 93%). A model which was trained with NBI images could only recognize NBI images in practice. However, a multicenter randomized controlled trial that compared a non-magnifying NBI with WLI indicated no significant difference in gastric cancer detection (65). Although NBI is currently regarded as the most broadly applied image-enhanced modality in AI research, the impact of other imaging modalities, such as the lately available linked-color imaging or blue-laser imaging modalities, need more studies for verification. For the number of training images, it seemed that the more images the machine trained, the more accurate the AI detection would be. The concept that a large number of images are a prerequisite to structure a learning model was also certified in the research conducted by Seguí et al. (66) for motility movement classification in wireless capsule endoscopy. A recent meta-analysis similarly indicated that a ten-fold increase in training data size could improve the accuracy of AI detection by 3% (67). Neural networks have the potential capacity for clinical practice and can be significantly popularized in the gastrointestinal field. However, CNN detection is temporarily in the stage of research. This study also had some limitations. A limited number of available studies fit the inclusion criteria since the novel technology has just been developed in recent years. Thus, the subgroup results were not completely reliable due to the limited number of studies. All the included studies were retrospective, which may lead to selection bias of included images, particularly in the validation dataset. In addition, few studies provided a solution to multiple gastrointestinal abnormalities as comparison, while most studies only researched the detection of a single abnormality, including Barrett's esophagus, Helicobacter pylori infection, early gastric cancer, atrophic gastritis, etc. (68–70), which is insufficient for clinical application. Moreover, an AI EGC detection model based on full-length videos was scarce, which postpones its general application in clinical practice. To overcome these limitations, several projects can be carried out in the future. More prospective studies can be designed for strict images, including criteria, high-definition image extraction and expert endoscopist involvement to prove higher level evidence. Luo et al. (71) has carried out a multicenter, case-control, prospective real-time diagnostic study on artificial intelligence for detection of esophagus and gastric cancer with accuracy of 0.955 (95% CI 0.952–0.957). GRAIDS algorithm, which was based on the concept of DeepLab's V3+ (72, 73), was utilized in this prospective study. Expanding the training image number is necessary to improve the machine recognition ability. On the other hand, the validation images are supposed to be larger. Training images extracted from different endoscopy modalities still need to be investigated to establish a popularized dataset. Currently, limited data have shown that the VGG-16, SSD, and SVM classifier models are credible computer-aided diagnosis algorithms. Another branch of deep learning, deep reinforcement learning (DRL), recently performed at the top level in the GO game in 2016 (74). DRL is likely to be applied in the EGC detection field. DRL combines deep learning with reinforcement learning, incorporating not only the excellent perception and distinguishing abilities of deep learning in visual tasks but also the decision-making capabilities of reinforcement learning (75). DRL has performed well in dealing with dynamic decision problems (74–76). However, DRL has not yet been used in clinical trials. Wu et al. (77) reported that the application of WISENSE, a mechanism that utilizes aspects of both CNN and DRL, could decrease the number of blind spots during an upper endoscopy, initially achieving an accuracy of 90.02%. The exploration of accurate algorithms is worthy of being explored.

Conclusion

This is the first meta-analysis to summarize current evidence of AI applications in EGC diagnosis. The AI applications seemed to be more accurate in parts of EGC detection than the endoscopists. The VGG-16, SSD, and SVM classifier models probably performed better according to the limited studies. When the number of training images is expanded, the accuracy will be improved. More strictly designed perspective studies with different reliable CNN algorithms are needed to make AI universal in clinical practice.

Data Availability Statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article/Supplementary Material.

Author Contributions

KJ: study concept and design and analysis and interpretation of data. XJ and YW: acquisition of data. SW, KN, ZZ, SJ, and PLiu: literature search. XJ, JP, and YH: figure processing. PLi and SL: critical revision of the manuscript for important intellectual content. FL: study supervision. All authors contributed to the article and approved the submitted version.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
  57 in total

1.  Modeling a MEMS deformable mirror using non-parametric estimation techniques.

Authors:  Dani Guzmán; Francisco Javier de Cos Juez; Richard Myers; Andrés Guesalaga; Fernando Sánchez Lasheras
Journal:  Opt Express       Date:  2010-09-27       Impact factor: 3.894

Review 2.  Deep learning.

Authors:  Yann LeCun; Yoshua Bengio; Geoffrey Hinton
Journal:  Nature       Date:  2015-05-28       Impact factor: 49.962

Review 3.  Performance measures for upper gastrointestinal endoscopy: a European Society of Gastrointestinal Endoscopy (ESGE) Quality Improvement Initiative.

Authors:  Raf Bisschops; Miguel Areia; Emmanuel Coron; Daniela Dobru; Bernd Kaskas; Roman Kuvaev; Oliver Pech; Krish Ragunath; Bas Weusten; Pietro Familiari; Dirk Domagk; Roland Valori; Michal F Kaminski; Cristiano Spada; Michael Bretthauer; Cathy Bennett; Carlo Senore; Mário Dinis-Ribeiro; Matthew D Rutter
Journal:  Endoscopy       Date:  2016-08-22       Impact factor: 10.093

4.  A multicenter randomized comparison between high-definition white light endoscopy and narrow band imaging for detection of gastric lesions.

Authors:  Tiing Leong Ang; Rapat Pittayanon; James Yun Wong Lau; Rungsun Rerknimitr; Shiaw Hooi Ho; Rajvinder Singh; Andrew Boon Eu Kwek; Daphne Shih Wen Ang; Philip Wai Yan Chiu; Sally Luk; Khean Lee Goh; Jeannie Peng Lan Ong; Jessica Yi-Lyn Tan; Eng Kiong Teo; Kwong Ming Fock
Journal:  Eur J Gastroenterol Hepatol       Date:  2015-12       Impact factor: 2.566

5.  Convolutional neural network classifier for distinguishing Barrett's esophagus and neoplasia endomicroscopy images.

Authors: 
Journal:  Conf Proc IEEE Eng Med Biol Soc       Date:  2017-07

6.  Real-time artificial intelligence for detection of upper gastrointestinal cancer by endoscopy: a multicentre, case-control, diagnostic study.

Authors:  Huiyan Luo; Guoliang Xu; Chaofeng Li; Longjun He; Linna Luo; Zixian Wang; Bingzhong Jing; Yishu Deng; Ying Jin; Yin Li; Bin Li; Wencheng Tan; Caisheng He; Sharvesh Raj Seeruttun; Qiubao Wu; Jun Huang; De-Wang Huang; Bin Chen; Shao-Bin Lin; Qin-Ming Chen; Chu-Ming Yuan; Hai-Xin Chen; Heng-Ying Pu; Feng Zhou; Yun He; Rui-Hua Xu
Journal:  Lancet Oncol       Date:  2019-10-04       Impact factor: 41.316

7.  A deep neural network improves endoscopic detection of early gastric cancer without blind spots.

Authors:  Lianlian Wu; Wei Zhou; Xinyue Wan; Jun Zhang; Lei Shen; Shan Hu; Qianshan Ding; Ganggang Mu; Anning Yin; Xu Huang; Jun Liu; Xiaoda Jiang; Zhengqiang Wang; Yunchao Deng; Mei Liu; Rong Lin; Tingsheng Ling; Peng Li; Qi Wu; Peng Jin; Jie Chen; Honggang Yu
Journal:  Endoscopy       Date:  2019-03-12       Impact factor: 10.093

Review 8.  Deep learning for wireless capsule endoscopy: a systematic review and meta-analysis.

Authors:  Shelly Soffer; Eyal Klang; Orit Shimon; Noy Nachmias; Rami Eliakim; Shomron Ben-Horin; Uri Kopylov; Yiftach Barash
Journal:  Gastrointest Endosc       Date:  2020-04-22       Impact factor: 9.427

9.  A Lesion-Based Convolutional Neural Network Improves Endoscopic Detection and Depth Prediction of Early Gastric Cancer.

Authors:  Hong Jin Yoon; Seunghyup Kim; Jie-Hyun Kim; Ji-Soo Keum; Sang-Il Oh; Junik Jo; Jaeyoung Chun; Young Hoon Youn; Hyojin Park; In Gyu Kwon; Seung Ho Choi; Sung Hoon Noh
Journal:  J Clin Med       Date:  2019-08-26       Impact factor: 4.241

10.  Application of artificial intelligence using a convolutional neural network for diagnosis of early gastric cancer based on magnifying endoscopy with narrow-band imaging.

Authors:  Hiroya Ueyama; Yusuke Kato; Yoichi Akazawa; Noboru Yatagai; Hiroyuki Komori; Tsutomu Takeda; Kohei Matsumoto; Kumiko Ueda; Kenshi Matsumoto; Mariko Hojo; Takashi Yao; Akihito Nagahara; Tomohiro Tada
Journal:  J Gastroenterol Hepatol       Date:  2020-07-28       Impact factor: 4.029

View more
  4 in total

1.  Objective Methods of 5-Aminolevulinic Acid-Based Endoscopic Photodynamic Diagnosis Using Artificial Intelligence for Identification of Gastric Tumors.

Authors:  Taro Yamashita; Hiroki Kurumi; Masashi Fujii; Takuki Sakaguchi; Takeshi Hashimoto; Hidehito Kinoshita; Tsutomu Kanda; Takumi Onoyama; Yuichiro Ikebuchi; Akira Yoshida; Koichiro Kawaguchi; Kazuo Yashima; Hajime Isomoto
Journal:  J Clin Med       Date:  2022-05-27       Impact factor: 4.964

Review 2.  Endoscopic Classifications of Early Gastric Cancer: A Literature Review.

Authors:  Mary Raina Angeli Fujiyoshi; Haruhiro Inoue; Yusuke Fujiyoshi; Yohei Nishikawa; Akiko Toshimori; Yuto Shimamura; Mayo Tanabe; Haruo Ikeda; Manabu Onimaru
Journal:  Cancers (Basel)       Date:  2021-12-26       Impact factor: 6.639

3.  Evaluation of Epidermal Growth Factor Receptor 2 Status in Gastric Cancer by CT-Based Deep Learning Radiomics Nomogram.

Authors:  Xiao Guan; Na Lu; Jianping Zhang
Journal:  Front Oncol       Date:  2022-07-11       Impact factor: 5.738

Review 4.  Advances in the Aetiology & Endoscopic Detection and Management of Early Gastric Cancer.

Authors:  Darina Kohoutova; Matthew Banks; Jan Bures
Journal:  Cancers (Basel)       Date:  2021-12-13       Impact factor: 6.639

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.