Literature DB >> 28580415

KID Project: an internet-based digital video atlas of capsule endoscopy for research purposes.

Anastasios Koulaouzidis1, Dimitris K Iakovidis2, Diana E Yung1, Emanuele Rondonotti3, Uri Kopylov4, John N Plevris1, Ervin Toth5, Abraham Eliakim4, Gabrielle Wurm Johansson5, Wojciech Marlicz6, Georgios Mavrogenis7, Artur Nemeth5, Henrik Thorlacius8, Gian Eugenio Tontini9.   

Abstract

BACKGROUND AND AIMS: Capsule endoscopy (CE) has revolutionized small-bowel (SB) investigation. Computational methods can enhance diagnostic yield (DY); however, incorporating machine learning algorithms (MLAs) into CE reading is difficult as large amounts of image annotations are required for training. Current databases lack graphic annotations of pathologies and cannot be used. A novel database, KID, aims to provide a reference for research and development of medical decision support systems (MDSS) for CE.
METHODS: Open-source software was used for the KID database. Clinicians contribute anonymized, annotated CE images and videos. Graphic annotations are supported by an open-access annotation tool (Ratsnake). We detail an experiment based on the KID database, examining differences in SB lesion measurement between human readers and a MLA. The Jaccard Index (JI) was used to evaluate similarity between annotations by the MLA and human readers.
RESULTS: The MLA performed best in measuring lymphangiectasias with a JI of 81 ± 6 %. The other lesion types were: angioectasias (JI 64 ± 11 %), aphthae (JI 64 ± 8 %), chylous cysts (JI 70 ± 14 %), polypoid lesions (JI 75 ± 21 %), and ulcers (JI 56 ± 9 %).
CONCLUSION: MLA can perform as well as human readers in the measurement of SB angioectasias in white light (WL). Automated lesion measurement is therefore feasible. KID is currently the only open-source CE database developed specifically to aid development of MDSS. Our experiment demonstrates this potential.

Entities:  

Year:  2017        PMID: 28580415      PMCID: PMC5452962          DOI: 10.1055/s-0043-105488

Source DB:  PubMed          Journal:  Endosc Int Open        ISSN: 2196-9736


Introduction

Capsule endoscopy (CE) has changed the field of small-bowel (SB) investigation 1 with the potential to become a panenteric diagnostic tool 2 . Computational methods incorporated into CE reading software can enhance diagnostic yield (DY) 3 . Several information technology (IT) groups have proposed software for detection of SB lesions/bleeding, reducing reading time, lesion localization, motility assessment, video enhancement and/or data management 1 3 . Reducing reading time is beneficial, especially in high volume centers. Previous work has shown that readers’ experience does not improve detection of lesions in CE 4 . Therefore, computer-aided detection/diagnosis (CAD) can improve DY. Despite prolific IT research, incorporating artificial intelligence (AI) systems into CE reading remains difficult 3 . The backbone of AI system development is based on machine learning algorithms (MLAs) for automatic detection, localization, and recognition of pathology in CE images and videos. A large amount of data, in the form of annotations, is required to train MLAs. Semantic annotations describe the content of CE videos and images, whereas graphic annotations are pixel-level labels indicating regions of interest (ROIs) (  Fig.1 ). Although there are some online databases 5 , these usually include the necessary semantic annotations, but lack graphic annotations of ROIs. Therefore, such material cannot be directly used by IT scientists for intelligent systems’ training or as a reference for their evaluation.
Fig. 1

 Dataset of angioectasia images and their corresponding graphic annotations, seen within the KID website interface.

Dataset of angioectasia images and their corresponding graphic annotations, seen within the KID website interface. A limited number of datasets composed of images with graphic annotations have become available in the context of IT studies 3 6 . A novel database, KID (κάψουλα interactive database; based on Greek for “capsule”) ( http://is-innovation.eu/kid/ ) was developed to fill this gap. It is available online, upon free registration, aiming to provide a reference for research on the development of medical decision support systems (MDSS) for CE, including the study of the performance of human observers in comparison to others and CAD.

Methods

Database

Open-source database (Oracle MySQL; https://www.mysql.com/ ) and web-gallery development software (Coppermine; http://coppermine-gallery.net/ ) were used. Software tools for video manipulation and image annotation were added to the KID website. To date, six centers (the KID working group) have contributed anonymized, annotated CE images/videos from various CE models; more than 2500 annotated CE images and 47 videos have been uploaded. These include images of (a) normal CE; (b) vascular lesions including angioectasias and/or bleeding; (c) inflammatory lesions, including mucosal aphthae and ulcers, erythema, cobblestoning, and luminal stenosis; (d) lymphangiectasias; and (e) polypoid lesions ( Fig. 2 ).
Fig. 2

 Top row, from left: P1 and P2 angioectasias, aphthae and ulcer, with corresponding graphic annotations made using Ratsnake beneath each image, showing the position, size and shape of the lesions in the images. Bottom row, from left: two images of nodular lymphangiectasias and two images of polypoid lesions, with graphic annotations below each image.

Top row, from left: P1 and P2 angioectasias, aphthae and ulcer, with corresponding graphic annotations made using Ratsnake beneath each image, showing the position, size and shape of the lesions in the images. Bottom row, from left: two images of nodular lymphangiectasias and two images of polypoid lesions, with graphic annotations below each image.

Image and video standards

Lesion categorization is based on the CE Structured Terminology (CEST) 7 . Contributions are of high quality (original resolution), not distorted by additional compression. For images, the recommended standard is ISO/IEC 15948 PNG (Portable Network Graphics), a popular platform-independent format with lossless compression. Other acceptable standards include: ISO/IEC, 14496-10, MPEG-4, AVC (Advanced Video Coding) and H.264. Supported formats for videos include F4V & FLV (Flash video).

Image annotation

The usefulness of KID relies on image annotations. Semantic and graphic annotations are supported by an open access, platform-independent annotation tool (Ratsnake) 8 . The graphic annotation process is shown in Fig. 3 and Video 1 . Semantic annotation is done through textual labels, and using standard web ontology language description logics (OWL DL) 9 . The quality of data and annotations submitted to KID are scrutinized by an international scientific committee ( http://is-innovation.eu/kid/committee.php ); contributions not meeting the aforementioned standards are rejected.
Fig. 3

 Use of the Ratsnake annotation tool to perform graphical annotation of an angioectasia on capsule endoscopy (CE). Left: original image. Right: graphic annotation of the angioectasia.

Use of the Ratsnake annotation tool to perform graphical annotation of an angioectasia on capsule endoscopy (CE). Left: original image. Right: graphic annotation of the angioectasia.

An experiment using the KID database: Computer-aided lesion size measurements based on color image segmentation

A total of 64 images of gastrointestinal lesions taken with MiroCam ® (IntroMedic Co., Seoul, Korea) were used. The lesions were: angioectasias (n = 27), lymphangiectasias (n = 9), ulcers (n = 9), chylous cysts (n = 8), polypoid lesions (n = 6), and small-bowel aphthae (n = 5). Graphic annotations made by expert readers (AK, ER, ET; > 2000 CE readings each) were used as lesion surface size reference standards. The images were automatically segmented into two regions: a ROI, i. e. the lesion in question, and the rest of the image. This was performed using the Localized Region-based Active Contour (LRAC) 10 algorithm, which is capable of segmenting regions characterized by heterogeneity in grayscale images; see Fig. 4 for a stepwise graphic presentation. The reader initializes the LRAC by defining a circular contour roughly on or around the lesion, starting at a random point in the image. The lesion did not need to be fully included in the initial contour. The algorithm calculates contours based on intensity histogram information (i. e. information on image brightness and intensity) from the regions inside and outside the contour. The calculations are performed locally, around each point along the contour. The algorithm continues to run until the overall similarity of the histograms inside and outside the contour is minimized. In this experiment, we extended the algorithm to the three components of the Commission internationale de l’éclairage-Lab (CIE-Lab) color space representation (instead of the standard RGB) 11 . Components of this space represent lightness ( L ), which is approximately equivalent to the respective grayscale image, quantity of red ( a  > 0) or quantity of green (–  a  > 0), quantity of yellow ( b  > 0) or quantity of blue (–  b  > 0) of a pixel ( Fig. 5 ). Fig. 6 shows the results of image segmentation using this algorithm applied to the a component of CIE-Lab, compared to in RGB. The Jaccard Index (JI) 12 was used to assess the similarity of the ROI obtained with the aid of LRAC compared to the graphically annotated ROI obtained by the expert readers (gold standard) per image, i. e. the agreement between the expert human readers and the algorithm. The JI is considered to be the most suitable and popular measure for the assessment of image segmentation algorithms 12 . It quantifies the overlap between two ROIs as the ratio of their intersection to their union with respect to the human readers. Therefore, it is independent from the measurement unit, e. g. pixels 2 or mm 2 , used to quantify the measured area. An illustrative example is provided in Fig. 7 .
Fig. 4

 Segmentation of image using the Localized Region-based Active Contour (LRAC) algorithm. a User-defined initial contour. b Contour deformation/morphing based on local histogram information on brightness and intensity in the various circular neighborhoods at each point on the contour. c Segmented image obtained.

Fig. 5

 CIE-Lab color wheel (left) compared to the RGB color wheel (right).

Fig. 6

 Image segmentation by Localized Region-based Active Contour (LRAC) algorithm. Top row, from left: original image of mucosal break with surrounding erythema; image segmentation using the a component of CIE-Lab; the final result of image segmentation where the contours have been defined and marked. Bottom row: the image when broken down into red (R), green (G) and blue (B) channels under the traditional RGB system.

Fig. 7 

Agreement between a human reader and the algorithm as quantified by the Jaccard Index (JI). Given a region annotated by a human expert (left) and a region annotated by the algorithm (right), the intersection of the two regions corresponds to the True Positive (TP) pixels, i. e. those actually belonging to the abnormality. The union of the two regions corresponds to the sum of the False Negative (FN), the False Positive (FP) and the TP. Thus, if the two regions perfectly coincide, FN = 0, FP = 0 and their intersection (TP) becomes equal to their union, resulting in JI = 100 %. If there is no match between the two regions, then TP = 0 and JI = 0.

Segmentation of image using the Localized Region-based Active Contour (LRAC) algorithm. a User-defined initial contour. b Contour deformation/morphing based on local histogram information on brightness and intensity in the various circular neighborhoods at each point on the contour. c Segmented image obtained. CIE-Lab color wheel (left) compared to the RGB color wheel (right). Image segmentation by Localized Region-based Active Contour (LRAC) algorithm. Top row, from left: original image of mucosal break with surrounding erythema; image segmentation using the a component of CIE-Lab; the final result of image segmentation where the contours have been defined and marked. Bottom row: the image when broken down into red (R), green (G) and blue (B) channels under the traditional RGB system. Agreement between a human reader and the algorithm as quantified by the Jaccard Index (JI). Given a region annotated by a human expert (left) and a region annotated by the algorithm (right), the intersection of the two regions corresponds to the True Positive (TP) pixels, i. e. those actually belonging to the abnormality. The union of the two regions corresponds to the sum of the False Negative (FN), the False Positive (FP) and the TP. Thus, if the two regions perfectly coincide, FN = 0, FP = 0 and their intersection (TP) becomes equal to their union, resulting in JI = 100 %. If there is no match between the two regions, then TP = 0 and JI = 0.

Results

The algorithm was evaluated for the measurement of six different types of small-bowel lesions, for each channel of CIE-Lab color space. The lesion areas were measured in pixel units, which, in the context of CE, is a more feasible and accurate approach. The average surface measurements closest to those performed by expert human readers were obtained by application of LRAC on the red-green scale of the CIE-Lab color space, with a JI of 67 ± 13 %. This result complements the findings in our previous study, indicating component a as an informative source of saliency for automated lesion detection 11 . The agreement between human readers and the algorithm per lesion type is summarized in   Table 1 . The most accurate measurements were obtained for lymphangiectasias, whereas this algorithm is less suitable for the measurement of ulcers.

Agreement between reviewers and software in measuring lesion size for various types of lesion seen on capsule endoscopy (CE).

Lesion JI, mean ± SD, %
Angioectasias64 ± 11
Aphthae64 ± 8
Chylous cysts70 ± 14
Lymphangiectasias81 ± 6
Polypoid lesions75 ± 21
Ulcers56 ± 9

Abbreviations: JI, Jaccard Index; SD, standard deviation.

Abbreviations: JI, Jaccard Index; SD, standard deviation.

Discussion

Human factors remain a barrier to timely and accurate CE diagnosis 4 . AI systems can improve clinical performance, patient safety, and resource utilization 1 3 . Open interdisciplinary exchange of information is key to technological advancement and therefore improved clinical outcomes 3 . New technological developments may not always meet pertinent healthcare needs due to little communication between software engineers and clinicians; furthermore, open access databases of endoscopic images are scarce, especially those specifically related to small-bowel CE 5 . This is despite growing clinical demand and use of CE as an investigative modality. However, such interactive formats are vital for engaging a new generation of clinicians; this is currently hindered by inadequately developed software 13 . Therefore, KID aims to be a comprehensive and all-encompassing resource for continuous development of CAD in CE, and to encourage two-way dialog between technological developers and end-users. For example, KID compiles images from all commercial CE models and is international, thus increasing its scope. The experiment detailed above shows that generally good agreement was achieved between expert human readers and the MLA in measuring the size of common small-bowel lesions. This implies automated lesion measurement is feasible, and MLAs could eventually replace or drastically reduce the workload of valuable human resources. In a recent study, van der Sommen et al. 14 detailed collaboration between IT engineers and clinicians to develop a CAD algorithm for diagnosis of early neoplasia in Barrett’s esophagus, with good results. An advantage of the method presented in this study over previous automated measurement approaches is its suitability for a variety of lesion types. In a recent study 15 using images of angiectasias available in KID, we showed that the interobserver agreement between CE reviewers, in terms of JI, in lesion annotation ranges between 65 ± 15 % and 67 ± 13 %, and the respective intraobserver agreement, between 69 ± 17 % and 71 ± 13 %. This dataset was similar in terms of the morphological characteristics of the displayed angiectasias, indicating that our MLA has a performance comparable to that of human readers. However, a limitation shown by the experiment is that it does not perform as well with all mucosal lesions. Further algorithm development is therefore required, showing the need for platforms such as KID. In conclusion, KID is, to our knowledge, the only database of CE images and videos with both graphic and semantic annotations, developed specifically for MDSS research. It provides a platform for data sharing and CAD software development. The experiments detailed are proof-of-principle studies demonstrating the potential for KID to fulfill this role.
  11 in total

1.  Detection of lesions during capsule endoscopy: physician performance is disappointing.

Authors:  YuanPu Zheng; Lauren Hawkins; Jordan Wolff; Olga Goloubeva; Eric Goldberg
Journal:  Am J Gastroenterol       Date:  2012-01-10       Impact factor: 10.864

2.  Supervised Evaluation of Image Segmentation and Object Proposal Techniques.

Authors:  Jordi Pont-Tuset; Ferran Marques
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2015-09-23       Impact factor: 6.226

Review 3.  Capsule endoscopy structured terminology (CEST): proposal of a standardized and structured terminology for reporting capsule endoscopy procedures.

Authors:  L Y Korman; M Delvaux; G Gay; F Hagenmuller; M Keuchel; S Friedman; M Weinstein; M Shetzline; D Cave; R de Franchis
Journal:  Endoscopy       Date:  2005-10       Impact factor: 10.093

Review 4.  Small-bowel capsule endoscopy: a ten-point contemporary review.

Authors:  Anastasios Koulaouzidis; Emanuele Rondonotti; Alexandros Karargyris
Journal:  World J Gastroenterol       Date:  2013-06-28       Impact factor: 5.742

5.  Capsule endoscopy as a panenteric diagnostic tool.

Authors:  M Hale; M E McAlindon
Journal:  Br J Surg       Date:  2013-11-19       Impact factor: 6.939

6.  Computer-aided detection of early neoplastic lesions in Barrett's esophagus.

Authors:  Fons van der Sommen; Svitlana Zinger; Wouter L Curvers; Raf Bisschops; Oliver Pech; Bas L A M Weusten; Jacques J G H M Bergman; Peter H N de With; Erik J Schoon
Journal:  Endoscopy       Date:  2016-04-21       Impact factor: 10.093

7.  Automatic lesion detection in capsule endoscopy based on color saliency: closer to an essential adjunct for reviewing software.

Authors:  Dimitris K Iakovidis; Anastasios Koulaouzidis
Journal:  Gastrointest Endosc       Date:  2014-08-01       Impact factor: 9.427

8.  Localizing region-based active contours.

Authors:  Shawn Lankton; Allen Tannenbaum
Journal:  IEEE Trans Image Process       Date:  2008-11       Impact factor: 10.856

9.  The informatics opportunities at the intersection of patient safety and clinical informatics.

Authors:  Peter M Kilbridge; David C Classen
Journal:  J Am Med Inform Assoc       Date:  2008-04-24       Impact factor: 4.497

10.  Ratsnake: a versatile image annotation tool with application to computer-aided diagnosis.

Authors:  D K Iakovidis; T Goudas; C Smailis; I Maglogiannis
Journal:  ScientificWorldJournal       Date:  2014-01-27
View more
  8 in total

1.  Towards a better understanding of annotation tools for medical imaging: a survey.

Authors:  Manar Aljabri; Manal AlAmir; Manal AlGhamdi; Mohamed Abdel-Mottaleb; Fernando Collado-Mesa
Journal:  Multimed Tools Appl       Date:  2022-03-25       Impact factor: 2.577

Review 2.  Surgical data science - from concepts toward clinical translation.

Authors:  Lena Maier-Hein; Matthias Eisenmann; Duygu Sarikaya; Keno März; Toby Collins; Anand Malpani; Johannes Fallert; Hubertus Feussner; Stamatia Giannarou; Pietro Mascagni; Hirenkumar Nakawala; Adrian Park; Carla Pugh; Danail Stoyanov; Swaroop S Vedula; Kevin Cleary; Gabor Fichtinger; Germain Forestier; Bernard Gibaud; Teodor Grantcharov; Makoto Hashizume; Doreen Heckmann-Nötzel; Hannes G Kenngott; Ron Kikinis; Lars Mündermann; Nassir Navab; Sinan Onogur; Tobias Roß; Raphael Sznitman; Russell H Taylor; Minu D Tizabi; Martin Wagner; Gregory D Hager; Thomas Neumuth; Nicolas Padoy; Justin Collins; Ines Gockel; Jan Goedeke; Daniel A Hashimoto; Luc Joyeux; Kyle Lam; Daniel R Leff; Amin Madani; Hani J Marcus; Ozanan Meireles; Alexander Seitel; Dogu Teber; Frank Ückert; Beat P Müller-Stich; Pierre Jannin; Stefanie Speidel
Journal:  Med Image Anal       Date:  2021-11-18       Impact factor: 13.828

3.  Development and Validation of an Artificial Intelligence Model for Small Bowel Capsule Endoscopy Video Review.

Authors:  Xia Xie; Yu-Feng Xiao; Xiao-Yan Zhao; Jian-Jun Li; Qiang-Qiang Yang; Xue Peng; Xu-Biao Nie; Jian-Yun Zhou; Yong-Bing Zhao; Huan Yang; Xi Liu; En Liu; Yu-Yang Chen; Yuan-Yuan Zhou; Chao-Qiang Fan; Jian-Ying Bai; Hui Lin; Anastasios Koulaouzidis; Shi-Ming Yang
Journal:  JAMA Netw Open       Date:  2022-07-01

4.  From labels to priors in capsule endoscopy: a prior guided approach for improving generalization with few labels.

Authors:  Anuja Vats; Ahmed Mohammed; Marius Pedersen
Journal:  Sci Rep       Date:  2022-09-20       Impact factor: 4.996

5.  Rhode Island gastroenterology video capsule endoscopy data set.

Authors:  Amber Charoen; Averill Guo; Panisara Fangsaard; Supakorn Taweechainaruemitr; Nuwee Wiwatwattana; Theekapun Charoenpong; Harlan G Rich
Journal:  Sci Data       Date:  2022-10-06       Impact factor: 8.501

6.  DINOSARC: Color Features Based on Selective Aggregation of Chromatic Image Components for Wireless Capsule Endoscopy.

Authors:  Michael D Vasilakakis; Dimitris K Iakovidis; Evaggelos Spyrou; Anastasios Koulaouzidis
Journal:  Comput Math Methods Med       Date:  2018-09-03       Impact factor: 2.238

7.  HyperKvasir, a comprehensive multi-class image and video dataset for gastrointestinal endoscopy.

Authors:  Hanna Borgli; Vajira Thambawita; Pia H Smedsrud; Steven Hicks; Debesh Jha; Sigrun L Eskeland; Kristin Ranheim Randel; Konstantin Pogorelov; Mathias Lux; Duc Tien Dang Nguyen; Dag Johansen; Carsten Griwodz; Håkon K Stensland; Enrique Garcia-Ceja; Peter T Schmidt; Hugo L Hammer; Michael A Riegler; Pål Halvorsen; Thomas de Lange
Journal:  Sci Data       Date:  2020-08-28       Impact factor: 6.444

Review 8.  Role of Artificial Intelligence in Video Capsule Endoscopy.

Authors:  Ioannis Tziortziotis; Faidon-Marios Laskaratos; Sergio Coda
Journal:  Diagnostics (Basel)       Date:  2021-06-30
  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.