| Literature DB >> 32038853 |
Zhang Xu1, Yu Tao1, Zheng Wenfang2,3, Lin Ne2,3, Huang Zhengxing1, Liu Jiquan1, Hu Weiling2,3, Duan Huilong1, Si Jianmin2,3.
Abstract
Esophagogastroduodenoscopy (EGD) has been widely applied for gastrointestinal (GI) examinations. However, there is a lack of mature technology to evaluate the quality of the EGD inspection process. In this Letter, the authors design a multi-task anatomy detection convolutional neural network (MT-AD-CNN) to evaluate the EGD inspection quality by combining the detection task of the upper digestive tract with ten anatomical structures and the classification task of informative video frames. The authors' model is able to eliminate non-informative frames of the gastroscopic videos and detect the anatomies in real time. Specifically, a sub-branch is added to the detection network to classify NBI images, informative and non-informative images. By doing so, the detected box will be only displayed on the informative frames, which can reduce the false-positive rate. They can determine the video frames on which each anatomical location is effectively examined, so that they can analyse the diagnosis quality. Their method reaches the performance of 93.74% mean average precision for the detection task and 98.77% accuracy for the classification task. Their model can reflect the detailed circumstance of the gastroscopy examination process, which shows application potential in improving the quality of examinations.Entities:
Keywords: EGD inspection process; EGD inspection quality; MT-AD-CNN; anatomies; authors design; biological organs; biomedical optical imaging; classification task; detected box; detection network; diagnosis quality; endoscopes; gastrointestinal examinations; gastroscopic videos; gastroscopy examination process; image classification; informative frames; informative video frames; inspection; learning (artificial intelligence); medical image processing; multitask anatomy detection convolutional neural network; multitask convolutional neural networks; neural nets; noninformative frames; noninformative images; patient diagnosis; upper digestive tract; upper gastrointestinal anatomy detection
Year: 2019 PMID: 32038853 PMCID: PMC6945683 DOI: 10.1049/htl.2019.0066
Source DB: PubMed Journal: Healthc Technol Lett ISSN: 2053-3713
Fig. 1Workflow of the proposed method
Fig. 2Examples of the labelled images for each class
Fig. 3Difference between MTL and the conventional approach
Fig. 4Architecture of the MT-AD
Dataset of the detection task
| Training | Testing | |
|---|---|---|
| patients | 2932 | 843 |
| images | 47,623 | 12,600 |
| labelled boxes | 59,513 | 15,762 |
Dataset of the classification task
| Training | Testing | |
|---|---|---|
| informative images | 10,053 | 2000 |
| non-informative images | 10,138 | 2000 |
| NBI images | 13,954 | 2000 |
| total images | 34,145 | 6000 |
Fig. 5P–R curves of the detection task
Average precision of each anatomy
| Method | mAP | Oesophagus | Cardia | Dentate line | Fundus | Body | Antrum | Angle | Pylorus | DB | DDP |
|---|---|---|---|---|---|---|---|---|---|---|---|
| MT-AD, % | 93.74 | 94.04 | 94.76 | 84.93 | 95.95 | 95.39 | 93.38 | 95.52 | 93.92 | 94.42 | 95.12 |
Fig. 6Heatmap of confusion matrix from the classification task
Fig. 7Examples of MT-AD's detection results
Fig. 8Statistical results from 83 gastric videos. Y-axis denotes the ratio of the valid time to the total time of the video (exclude NBI)
Fig. 9Statistical results from 83 gastric videos. Y-axis denotes the ratio of the time at which the specific anatomy is examined to the valid time of this video
Images for each class of the detection dataset
| Class | Training | Testing |
|---|---|---|
| oesophagus | 6851 | 1766 |
| dentate line | 4951 | 1356 |
| cardia | 3955 | 1054 |
| fundus | 5684 | 1503 |
| body | 10,723 | 2932 |
| antrum | 7209 | 1916 |
| angle | 5050 | 1357 |
| pylorus | 5930 | 1597 |
| DB | 4832 | 1175 |
| DDP | 4328 | 1106 |
| total | 59,513 | 15,762 |