| Literature DB >> 29056541 |
Satoki Shichijo1, Shuhei Nomura2, Kazuharu Aoyama3, Yoshitaka Nishikawa4, Motoi Miura5, Takahide Shinagawa6, Hirotoshi Takiyama6, Tetsuya Tanimoto7, Soichiro Ishihara8, Keigo Matsuo9, Tomohiro Tada6.
Abstract
BACKGROUND AND AIMS: The role of artificial intelligence in the diagnosis of Helicobacter pylori gastritis based on endoscopic images has not been evaluated. We constructed a convolutional neural network (CNN), and evaluated its ability to diagnose H. pylori infection.Entities:
Keywords: Artificial intelligence; Convolutional neural networks; Endoscopy; Helicobacter pylori
Mesh:
Year: 2017 PMID: 29056541 PMCID: PMC5704071 DOI: 10.1016/j.ebiom.2017.10.014
Source DB: PubMed Journal: EBioMedicine ISSN: 2352-3964 Impact factor: 8.143
Fig. 1Representative endoscopic images of Helicobacter pylori-positive, and –negative stomach. Atrophy and diffuse redness are seen in the presence of infection. A regular arrangement of collecting venules (RAC) is seen in the uninfected stomach.
Baseline characteristics.
| Characteristics | Development data set | Test data set |
|---|---|---|
| No. of images | 32,208 | 11,481 |
| No. of endoscopists | 33 | 13 |
| No. of patients | 1768 | 397 |
| Age, mean (SD), y | 52.7 (13.2) | 50.4 (11.2) |
| Sex, No. (%) | ||
| Male | 480 (45) | 168 (43) |
| Female | 598 (55) | 226 (57) |
| Positive | 753 (43) | 72 (18) |
| Negative | 1015 (57) | 325 (82) |
SD, standard deviation.
Data were available for 1078 cases.
Fig. 2Patient recruitment flowchart.
Fig. 3Deep convolutional neural network (CNN) layout.
We used a CNN technique for image classification. Data flow is from bottom to top direction. With a given input image, the CNN architecture produces a probability distribution over classes as H. pylori positive or negative. The GoogLeNet, a deep CNN of 22 layers, is pre-trained on the ImageNet dataset and fine-tuned on our own dataset of about 400,000 endoscopic images, which are pre-augmented.
GoogLeNet architecture published from https://arxiv.org/abs/1409.4842.
Fig. 5Receiver operating curves for CNN trained by categorized data.
The CNN output demonstrates better probability following a training based on location-based classification of images. The area under the receiver operating curve is now 93%. Each endoscopist's prediction is represented by a single red point. The green point is the average prediction of the endoscopists.
Diagnostic accuracy: CNN vs. endoscopists.
| CNN | Endoscopists | |||||
|---|---|---|---|---|---|---|
| First CNN | Secondary CNN | Certified | Relatively experienced | Beginner | Total | |
| No. of endoscopists | 6 | 9 | 8 | 23 | ||
| Sensitivity (SD), % | 81.9 | 88.9 | 85.2 (4.5) | 81.0 (10.2) | 72.2 (14.3) | 79.0 (11.7) |
| Specificity (SD), % | 83.4 | 87.4 | 89.3 (2.6) | 85.1 (8.7) | 76.3 (10.8) | 83.2 (9.8) |
| Accuracy (SD), % | 83.1 | 87.7 | 88.9 (2.9) | 84.4 (7.1) | 75.6 (8.2) | 82.4 (8.4) |
| AUC | 0.89 | 0.93 | ||||
| Time (SD), min | 3.3 | 3.2 | 252.5 (92.3) | 236.1 (51.9) | 206.6 (54.7) | 230.1 (65.0) |
SD, standard deviation; AUC, area under the receiver operating curve.
Fig. 4Receiver operating curves for CNN trained by uncategorized data and prediction of the endoscopists.
Each endoscopist's prediction is represented by a single red point. The green point is the average prediction of the endoscopists. The CNN outputs a H. pylori probability P per image, and then the program calculates a mean square of the probabilities per patient. The area under the receiver operating curve is over 89%.