Literature DB >> 35511456

Development and evaluation of a deep learning model to improve the usability of polyp detection systems during interventions.

Markus Brand1, Joel Troya1, Adrian Krenzer1,2, Zita Saßmannshausen1, Wolfram G Zoller3, Alexander Meining1, Thomas J Lux1, Alexander Hann1.   

Abstract

BACKGROUND: The efficiency of artificial intelligence as computer-aided detection (CADe) systems for colorectal polyps has been demonstrated in several randomized trials. However, CADe systems generate many distracting detections, especially during interventions such as polypectomies. Those distracting CADe detections are often induced by the introduction of snares or biopsy forceps as the systems have not been trained for such situations. In addition, there are a significant number of non-false but not relevant detections, since the polyp has already been previously detected. All these detections have the potential to disturb the examiner's work.
OBJECTIVES: Development and evaluation of a convolutional neuronal network that recognizes instruments in the endoscopic image, suppresses distracting CADe detections, and reliably detects endoscopic interventions.
METHODS: A total of 580 different examination videos from 9 different centers using 4 different processor types were screened for instruments and represented the training dataset (519,856 images in total, 144,217 contained a visible instrument). The test dataset included 10 full-colonoscopy videos that were analyzed for the recognition of visible instruments and detections by a commercially available CADe system (GI Genius, Medtronic).
RESULTS: The test dataset contained 153,623 images, 8.84% of those presented visible instruments (12 interventions, 19 instruments used). The convolutional neuronal network reached an overall accuracy in the detection of visible instruments of 98.59%. Sensitivity and specificity were 98.55% and 98.92%, respectively. A mean of 462.8 frames containing distracting CADe detections per colonoscopy were avoided using the convolutional neuronal network. This accounted for 95.6% of all distracting CADe detections.
CONCLUSIONS: Detection of endoscopic instruments in colonoscopy using artificial intelligence technology is reliable and achieves high sensitivity and specificity. Accordingly, the new convolutional neuronal network could be used to reduce distracting CADe detections during endoscopic procedures. Thus, our study demonstrates the great potential of artificial intelligence technology beyond mucosal assessment.
© 2022 The Authors. United European Gastroenterology Journal published by Wiley Periodicals LLC on behalf of United European Gastroenterology.

Entities:  

Keywords:  CADe; colonoscopy; deep learning; instrument; intervention

Mesh:

Year:  2022        PMID: 35511456      PMCID: PMC9189459          DOI: 10.1002/ueg2.12235

Source DB:  PubMed          Journal:  United European Gastroenterol J        ISSN: 2050-6406            Impact factor:   6.866


Summarize the established knowledge on this subject Multiple computer‐aided diagnosis (CADe) systems for polyp detection are currently introduced into clinical practice. CADe results in multiple distracting detections, especially during therapeutic interventions when instruments are visible. What are the significant and/or new findings of this study? Development and evaluation of a deep learning model to recognize visible instruments that are used for therapeutic intervention in gastrointestinal endoscopy. Our model automatically prevents distracting CADe detections during therapeutic interventions.

INTRODUCTION

Artificial intelligence (AI) for colonic polyp detection is the most important application of this new technology in gastrointestinal endoscopy to date. Efficiency and functionality of these computer‐aided detection (CADe) systems have been demonstrated in several randomized trials. , , , , , However, CADe systems also show many false positive (FP) detections. , These false markings can affect the examiner's concentration. If a false detection occurs in addition to a relevant finding, the examiner's attention may be distracted, leading to missed findings in the worst case. Daily use of CADe systems shows that endoscopic interventions (especially biopsies and polypectomies) lead to many false activations of CADe systems. In this case, false positive activations occur due to the inserted instruments (forceps, needle, snare), but also due to intervention on the mucosa itself (injection, resection, clipping). In addition, there are a significant number of non‐false but not relevant detections, since the polyp has already been previously detected. To enable the investigators to put their full concentration on the intervention, no distracting AI signals regarding polyp detection should be visible during the procedure. Therefore, the aim of the current study was to develop and evaluate an AI system that reliably detects introduced instruments in order to disable the CADe system during an intervention and avoid distracting detections.

METHODS

Training dataset

Data from nine different center in Germany, two university hospitals, one community‐based hospital and six gastroenterology practices, were collected retrospectively from March 2019 to August 2021. A total of 519,856 images were selected from 580 randomly selected different colonoscopy videos for building the training dataset. Of all images in the training dataset, 144,217 (27.7%) contained a visible instrument (Figure 1). The types of instruments used for training the model included graspers, hot and cold snares, injection needles and clips. No minimum or maximum number of images per instrument in a colonoscopy was predefined for training the model. Images of good and poor quality (e.g., blurry images) were chosen for model training in order to represent a real‐life scenario. The colonoscopies were performed using different processors including CV‐190 and CV‐170 (Olympus Europa SE & Co. KG, Hamburg, Germany), Image1 S (Karl Storz SE & Co. KG, Tuttlingen, Germany) and EPK‐i7000 (Pentax Europe GmbH, Hamburg, Germany) and were recorded using a standard computer with a video grabber (DeckLink Mini Recorder, Blackmagic Design Pty Ltd., Melbourne, Australia) and a custom recording software. Representative images of the four different processor types are displayed in Supplementary Figure 1.
FIGURE 1

Characteristics of the training dataset containing images with and without visible instruments captured using the four different endoscopy processor types

Characteristics of the training dataset containing images with and without visible instruments captured using the four different endoscopy processor types To reduce almost identical images in the dataset, images from the same colonoscopy were filtered to exclude neighboring images. For training the convolutional neuronal network (CNN), the dataset was split into a train (90%) and a validation (10%) dataset. To prevent bias, all images of one colonoscopy were either included in the train or the validation dataset.

Preprocessing and CNN training

Initially, a region of interest for each used processor type was defined and images were cropped accordingly. Afterward, images were zero padded and resized to a dimension of 512 × 512 pixels to yield uniform images. For train data, the image augmentation pipeline (Supplementary Code Section 1) was applied. All images underwent the standard procedure of image normalization (Supplementary Code Section 2), so that each color and brightness value are standardized. Resulting constant values for this were calculated on all train data images and were always used to normalize each input image. CNN training is described in detail in Supplementary Material. The current instrument detection software is freely available to download for research purposes (https://github.com/Maddonix/instrument_detection).

Model testing

The CNN for detecting visible instruments was tested in the withdrawal phase of a set of 10 full‐length colonoscopy videos by analyzing its performance in each single image of the videos. To stabilize the prediction results, a running mean function was applied. This was performed to avoid erroneous suppressions of AI detections caused by our CNN. Here, we assigned the current video image the majority label of itself as well as the previous 14 video images, representing a threshold of 467 ms. This same dataset was used to test the change in performance of a CADe system (GI Genius, Medtronic Inc., Ireland, Version March 2020) in the reduction of distracting activations with the developed instrument detection system.

Ethics approval

Patients provided written informed consent prior to video recording. The ethics committee of the University hospital Würzburg approved retrospective analysis of the data used in this study.

Statistical analysis

Two evaluations were statistically analyzed: the capabilities of the instrument detection system and the reduction of CADe distracting activations. Per‐frame sensitivity and specificity, accuracy, and Receiver Operating Characteristic (ROC) were calculated for both evaluations. Sensitivity has been defined as the ratio between the number of frames with a visible instrument that were correctly detected (TP) and the total number of frames with a visible instrument (TP+FN). Specificity was defined as the ratio between the number of frames without a visible instrument that were correctly assessed (TN) and the sum of the total number of frames with a false detection and the TN frames (FP+TN). Accuracy was defined as the ratio between the number of correct system assessments (TP+TN) and the total number of frames. Metrics where weighted average to compensate for the imbalance of images with/out a visible instrument. For the calculation of the weighted average metrics the parameter “average” in every used function of the sklearn.metrics module from scikit‐learn 1.0.2 package was set to “weighted”. All calculations were performed using Python Software (version 3.6).

RESULTS

Characteristics of the patient cohort

The test dataset comprised 10 full‐length colonoscopy videos from 10 different patients. Men and women were equally represented, the mean age was 57.1 (interquartile range; 46–65) and the mean Boston Bowel Preparation Scale was 6.9 (range; 6–9) (Supplementary Table 1). The total duration of the withdrawal phase, with the duration of interventions included, was 1 h and 25 min, corresponding to 153,623 single video frames. During this time, instruments were visible for a total of 7 min and 12 s on the screen These 10 videos included a total of 12 different interventions, where 19 different endoscopic through the scope instruments were used: 4 cold snares, 11 graspers, 1 hot snare, 2 needles and 1 clip (Table 1).
TABLE 1

Characteristics and performance of the instrument detection system in the test dataset

InterventionType of instrumentNumber of visible instrument framesSensitivity (%)Specificity (%)Disturbing CADe activations (frames)Disturbing CADe activations avoided (frames)False‐avoided CADe activations (frames)Total number of CADe activations
Video 1 PolypectomySnare72898.6099.51377330133352
PolypectomySnare1262
Video 2 PolypectomyGrasper14299.7799.76660396
Video 3 PolypectomyGrasper26999.2199.6813710251252
Video 4 PolypectomyGrasper17498.8798.43882302
Video 5 PolypectomyNeedle40799.2299.6412321184547834
Snare931
Video 6 PolypectomyGrasper113699.3199.671611506531
Video 7 PolypectomySnare276098.0198.35741736501577
Video 8 PolypectomyNeedle249396.9096.58192319062047737
Hot snare1048
Clip292
Video 9 PolypectomySnare25599.3399.6210184211361
Video 10 Random biopsies5x Grasper75198.1498.9815312221099
PolypectomyGrasper871

Abbreviation: CADe, Computer‐aided detection system.

Characteristics and performance of the instrument detection system in the test dataset Abbreviation: CADe, Computer‐aided detection system.

Performance of the instrument detection system

The CNN overall accuracy achieved in the detection of visible instruments in the test dataset was 98.59%. Sensitivity and specificity were 98.55% and 98.92%, respectively. The grasper was the instrument that was best detected by the system, with a sensitivity of 99.08% and a specificity of 99.36%, whereas the snare, with a sensitivity of 98.21% and a specificity of 98.51%, was the most difficult instrument to detect, probably because often only the wire was visible. Representative images of a grasper, a snare and a false positive detection of the CNN with the corresponding heat map that depicts the image areas that are recognized as an instrument are presented in Figure 2. The ROC curve illustrating the diagnostic ability of our instrument detection system is depicted in Figure 3. No marked differences in performance were observed relating to BBPS value, that ranged from 6 to 9 (Supplementary Table 1).
FIGURE 2

Grasper (upper row), snare (middle row), and a false positive detection (lower row) of the instrument detecting CNN with the corresponding gradient‐weighted class activation mapping (Grad‐CAM). Grad‐CAM images on the right side visualize areas responsible for the CNN prediction as an instrument

FIGURE 3

Receiver Operating Characteristic Curve of the instrument detection CNN visualizes specificity. While adjusting classification thresholds, the TP rate reaches 96.58% while maintaining a FP rate of 1% resulting in an area under the curve of 0.9971. CNN, convolutional neuronal network; FP, false positive; TP, true positive

Grasper (upper row), snare (middle row), and a false positive detection (lower row) of the instrument detecting CNN with the corresponding gradient‐weighted class activation mapping (Grad‐CAM). Grad‐CAM images on the right side visualize areas responsible for the CNN prediction as an instrument Receiver Operating Characteristic Curve of the instrument detection CNN visualizes specificity. While adjusting classification thresholds, the TP rate reaches 96.58% while maintaining a FP rate of 1% resulting in an area under the curve of 0.9971. CNN, convolutional neuronal network; FP, false positive; TP, true positive

Reduction of CADe distracting activations

A total of 25,441 activations were triggered by the CADe system in the test dataset. These activations included detected polyps and false positive detections. 4839 activations (19.02%) occurred when an instrument was visible and were regarded as distracting CADe activations. Especially significant was the amount of distracting activation caused by snares or needles. Our system was able to avoid 4628 of these activations, representing a sensitivity value of 95.64%. Regarding the number of CADe activations that were falsely avoided, the value in frames amounts to 357. Out of those, 292 contained polyps that were previously detected. This implies that the system has a specificity of 98.62% in terms of performance in preventing distracting CADe activations. The overall accuracy was of 99.17%. The metrics of the developed instrument detection system per intervention and its performance in preventing distracting CADe activations are presented in Table 1 and exemplarily illustrated in Figure 4. In addition, Figure 5 and Video S1 present a graphical example of how the AI system works.
FIGURE 4

Schematic overview of the images with (red) and without (blue) visible instruments in a coloscopy video. The first row represents the manual annotations of whether the corresponding image contains a visible instrument. The second row represents the predictions output by our CNN. The third row represents the distracting CADe activations successfully prevented (green) or unsuccessfully prevented (yellow) by using the developed instrument detection CNN. The inset shows 160 frames (one dot per frame) which correlate to 5.33 s in the video. CADe, Computer‐aided detection system; CNN, convolutional neuronal network

FIGURE 5

Single images of a polypectomy involving a needle for submucosal injection (upper row) and a snare (lower row) using the computer‐aided polyp detection system (CADe) (left) and the additional CADe preventing instrument detection system (right). Video S1: Head‐to‐head comparison of a colonoscopy video sequence with (right) and without (left) the use of the instrument detection convolutional neuronal network

Schematic overview of the images with (red) and without (blue) visible instruments in a coloscopy video. The first row represents the manual annotations of whether the corresponding image contains a visible instrument. The second row represents the predictions output by our CNN. The third row represents the distracting CADe activations successfully prevented (green) or unsuccessfully prevented (yellow) by using the developed instrument detection CNN. The inset shows 160 frames (one dot per frame) which correlate to 5.33 s in the video. CADe, Computer‐aided detection system; CNN, convolutional neuronal network Single images of a polypectomy involving a needle for submucosal injection (upper row) and a snare (lower row) using the computer‐aided polyp detection system (CADe) (left) and the additional CADe preventing instrument detection system (right). Video S1: Head‐to‐head comparison of a colonoscopy video sequence with (right) and without (left) the use of the instrument detection convolutional neuronal network

DISCUSSION

Since the introduction of commercially available AI‐systems for colorectal polyp detection, the use of these promising systems in daily practice is increasing. The great potential of AI‐systems is currently in the field of diagnostics, as CADe systems support the examiner in real time and with high sensitivity. , , Since CADe systems have been trained with diagnostic polyp images, they achieve high sensitivity for native polyps in the colon. However, changes to the mucosa in the course of an intervention (e.g., injection) lead to false positive detections, as the systems have not been trained for such situations. The instruments used during the intervention also lead to many false positive detections that may disturb the investigator's concentration. In addition, there are many non‐false but irrelevant detections because the polyp causing the intervention has been previously detected. To enable the investigators to put their full concentration on the intervention, no distracting AI signals should be visible during the intervention. This could be achieved by suppressing the CADe signal during the intervention, since polyp detection is not necessary during the intervention. Currently, the endoscopist can only manually turn off the CADe system and turn it back on after the procedure. Some systems require a button to be pressed on the processor, as not all systems can be controlled via a button on the endoscope. However, it is possible that the endoscopist forgets to turn the system back on after the procedures. Therefore, automatically stopping and starting the CADe system would increase the comfort for the endoscopist and prevent the CADe system from being accidentally switched off permanently. Our novel AI system detects inserted instruments with high sensitivity and specificity. Therefore, the system can capture the time frame of an endoscopic intervention with high accuracy. This would enable the suppression of the CADe signal for the duration of the intervention to focus the investigator's concentration on the intervention. The suppression relates not only to false positive detections but instead to all CADe detections during an intervention that do not add value to the endoscopic image. The requirements for such a tool detection system are very high, as suppression of the CADe signal outside of an intervention (false positive instrument detection) may increase the risk of missing other visible polyps. Our study shows that our new AI system achieves a very high specificity, which is sufficient for this purpose. To obtain this high specificity, our system was trained with a large number of images from multiple centers using different endoscopy processors. The number of training images we used is comparable to the number used in development of other CADe systems. , In addition, the optimized algorithm presents only a short delay of 467 ms, that allows for the real‐time use in combination with a CADe system. Since the sensitivity of our AI system is in a high range, the instruments introduced were missed in only a few frames during an intervention. This applies in particular to the insertion and removal of an instrument where only a small portion of it is visible at the edge of the endoscopic view. Once the instrument is in the normal working position, it is quickly and reliably detected by the AI system. Thus, the crucial part of the intervention is captured by our instrument detection system. However, a problem with instrument detection arises when an instrument is pressed so firmly into the mucosa that it is barely visible. In this situation, the instrument recognition works accordingly worse. Nevertheless, our video analysis showed that the new AI system significantly reduced the number of false‐positive CADe detections during an endoscopic intervention. While many publications on AI systems only use short, specially selected video sequences in the evaluation phase, our system was tested on full‐length colonoscopy, which brings the results much closer to the real examination situation. Interestingly, the commercially available CADe system seems to generate more detections when a snare is used in comparison to a grasper. There might be different explanations for this phenomenon. The first one is that the snare produces folds around the polyp by pressing the wire against the mucosa. Those folds are then falsely interpreted as polyps by the commercially available CADe system. Another explanation is that the total time that snares are visible in our dataset is longer than that of graspers. The longer visibility results in more CADe detections. Lastly, the not openly available training dataset of the commercially available CADe system might contain images of graspers and snares in an unbalanced manner. The implementation of our freely available instrument detection AI system in an existing CADe system could be done by controlling the input signal of the examination monitor. Here, the instrument detection AI would analyze the raw endoscopy processor output signal in parallel with the CADe system. During withdrawal with no visible instrument, the CADe signal would be displayed to the examiner. Thus, allowing the examiner to fully benefit from the high polyp detection rate of a commercially available CADe system. When an instrument is detected by the new AI, an automatic switch would display the raw processor signal instead. Alternatively, the AI system for instrument recognition can also be integrated as a filter directly into an existing or newly developed CADe system. By integrating our system in a single CADe system, the process would be more comfortable. To the best of our knowledge, an AI system for instrument detection using deep learning methods has not been developed in gastrointestinal endoscopy. Therefore, we present the first AI system in the field that enables recognition of endoscopic interventions by instrument detection. In addition to the mentioned application, the system could also be useful for automated recording of intervention times or withdrawal time. This could potentially help in obtaining objective data to assess the quality of colonoscopies. However, our study has several limitations. The new AI system was evaluated using previously recorded videos. Therefore, the mentioned implementation of the AI system in daily practice must be tested in future prospective studies to evaluate clinical benefit. Another possibility would be to investigate (e.g. by eye‐tracking) whether the examiner's attention could be better focused on the intervention by reducing distracting CADe signals. Other limitations include that, to facilitate the generation of a large‐annotated training dataset in a short period of time, no predefined protocol was used for video selection. A quantitative identification of the causes of false positive detections of our CNN need to be evaluated in future studies. CADe systems already achieved a remarkable benefit in randomized controlled trials. Future developments of those systems include improving usability by adding customizable features. The commercially available system that was used in our study for example, presents a well‐studied CADe function. Other systems incorporate computer‐aided diagnosis (CADx) that is only turned on when virtual chromoendoscopy is activated by the examiner. In the case of the ENDO‐AID CADe system by Olympus the examiner has the possibility to choose how many CADe detection boxes should be maximally displayed on the screen. The examiner can even choose from two different CADe modes that presumably present different sensitivities. In other words, in general, these devices are incorporating customizable modules that increase the usability and, therefore, the value of the device. Our work aligns in this direction by contributing to the prevention of distracting CADe activations during interventions. In conclusion, our study shows that instrument detection using AI technology is reliable and achieves high sensitivity and specificity. Therefore, the new AI system could be helpful to reduce distracting CADe detections during endoscopic procedures. Although the clinical benefit of the new AI system needs further evaluation, our study demonstrates the great potential of AI technology beyond mucosal assessment.

CONFLICT OF INTEREST

The authors have no conflicts of interest to declare. Supporting Information S1 Click here for additional data file. Video S1 Click here for additional data file.
  20 in total

1.  Effect of a deep-learning computer-aided detection system on adenoma detection during colonoscopy (CADe-DB trial): a double-blind randomised study.

Authors:  Pu Wang; Xiaogang Liu; Tyler M Berzin; Jeremy R Glissen Brown; Peixi Liu; Chao Zhou; Lei Lei; Liangping Li; Zhenzhen Guo; Shan Lei; Fei Xiong; Han Wang; Yan Song; Yan Pan; Guanyu Zhou
Journal:  Lancet Gastroenterol Hepatol       Date:  2020-01-22

Review 2.  Performance measures for lower gastrointestinal endoscopy: a European Society of Gastrointestinal Endoscopy (ESGE) quality improvement initiative.

Authors:  Michal F Kaminski; Siwan Thomas-Gibson; Marek Bugajski; Michael Bretthauer; Colin J Rees; Evelien Dekker; Geir Hoff; Rodrigo Jover; Stepan Suchanek; Monika Ferlitsch; John Anderson; Thomas Roesch; Rolf Hultcranz; Istvan Racz; Ernst J Kuipers; Kjetil Garborg; James E East; Maciej Rupinski; Birgitte Seip; Cathy Bennett; Carlo Senore; Silvia Minozzi; Raf Bisschops; Dirk Domagk; Roland Valori; Cristiano Spada; Cesare Hassan; Mario Dinis-Ribeiro; Matthew D Rutter
Journal:  United European Gastroenterol J       Date:  2017-03-16       Impact factor: 4.623

3.  Lower Adenoma Miss Rate of Computer-Aided Detection-Assisted Colonoscopy vs Routine White-Light Colonoscopy in a Prospective Tandem Study.

Authors:  Pu Wang; Peixi Liu; Jeremy R Glissen Brown; Tyler M Berzin; Guanyu Zhou; Shan Lei; Xiaogang Liu; Liangping Li; Xun Xiao
Journal:  Gastroenterology       Date:  2020-06-17       Impact factor: 22.682

Review 4.  Current status and limitations of artificial intelligence in colonoscopy.

Authors:  Alexander Hann; Joel Troya; Daniel Fitting
Journal:  United European Gastroenterol J       Date:  2021-06-07       Impact factor: 4.623

5.  Performance of a new integrated computer-assisted system (CADe/CADx) for detection and characterization of colorectal neoplasia.

Authors:  Jochen Weigt; Alessandro Repici; Giulio Antonelli; Ahmed Afifi; Leon Kliegis; Loredana Correale; Cesare Hassan; Helmut Neumann
Journal:  Endoscopy       Date:  2021-04-20       Impact factor: 9.776

6.  Benchmarking definitions of false-positive alerts during computer-aided polyp detection in colonoscopy.

Authors:  Erik A Holzwanger; Mohammad Bilal; Jeremy R Glissen Brown; Shailendra Singh; Aymeric Becq; Kenneth Ernest-Suarez; Tyler M Berzin
Journal:  Endoscopy       Date:  2021-01-18       Impact factor: 9.776

7.  Real-time automatic detection system increases colonoscopic polyp and adenoma detection rates: a prospective randomised controlled study.

Authors:  Pu Wang; Tyler M Berzin; Jeremy Romek Glissen Brown; Shishira Bharadwaj; Aymeric Becq; Xun Xiao; Peixi Liu; Liangping Li; Yan Song; Di Zhang; Yi Li; Guangre Xu; Mengtian Tu; Xiaogang Liu
Journal:  Gut       Date:  2019-02-27       Impact factor: 23.059

8.  Application of artificial intelligence using a convolutional neural network for diagnosis of early gastric cancer based on magnifying endoscopy with narrow-band imaging.

Authors:  Hiroya Ueyama; Yusuke Kato; Yoichi Akazawa; Noboru Yatagai; Hiroyuki Komori; Tsutomu Takeda; Kohei Matsumoto; Kumiko Ueda; Kenshi Matsumoto; Mariko Hojo; Takashi Yao; Akihito Nagahara; Tomohiro Tada
Journal:  J Gastroenterol Hepatol       Date:  2020-07-28       Impact factor: 4.029

9.  Development and evaluation of a deep learning model to improve the usability of polyp detection systems during interventions.

Authors:  Markus Brand; Joel Troya; Adrian Krenzer; Zita Saßmannshausen; Wolfram G Zoller; Alexander Meining; Thomas J Lux; Alexander Hann
Journal:  United European Gastroenterol J       Date:  2022-05-05       Impact factor: 6.866

10.  Real-time use of artificial intelligence in the evaluation of cancer in Barrett's oesophagus.

Authors:  Alanna Ebigbo; Robert Mendel; Andreas Probst; Johannes Manzeneder; Friederike Prinz; Luis A de Souza; Joao Papa; Christoph Palm; Helmut Messmann
Journal:  Gut       Date:  2019-09-20       Impact factor: 23.059

View more
  2 in total

1.  Fine-tuning the artificial intelligence experience in endoscopy.

Authors:  Keith Siau; Tyler M Berzin
Journal:  United European Gastroenterol J       Date:  2022-05-20       Impact factor: 6.866

2.  Development and evaluation of a deep learning model to improve the usability of polyp detection systems during interventions.

Authors:  Markus Brand; Joel Troya; Adrian Krenzer; Zita Saßmannshausen; Wolfram G Zoller; Alexander Meining; Thomas J Lux; Alexander Hann
Journal:  United European Gastroenterol J       Date:  2022-05-05       Impact factor: 6.866

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.