Literature DB >> 30304439

CANDI: an R package and Shiny app for annotating radiographs and evaluating computer-aided diagnosis.

Marcus A Badgeley1,2,3, Manway Liu3, Benjamin S Glicksberg4, Mark Shervey1,2, John Zech5, Khader Shameer6, Joseph Lehar7, Eric K Oermann8, Michael V McConnell3,9, Thomas M Snyder3, Joel T Dudley1,2.   

Abstract

MOTIVATION: Radiologists have used algorithms for Computer-Aided Diagnosis (CAD) for decades. These algorithms use machine learning with engineered features, and there have been mixed findings on whether they improve radiologists' interpretations. Deep learning offers superior performance but requires more training data and has not been evaluated in joint algorithm-radiologist decision systems.
RESULTS: We developed the Computer-Aided Note and Diagnosis Interface (CANDI) for collaboratively annotating radiographs and evaluating how algorithms alter human interpretation. The annotation app collects classification, segmentation, and image captioning training data, and the evaluation app randomizes the availability of CAD tools to facilitate clinical trials on radiologist enhancement.
AVAILABILITY AND IMPLEMENTATION: Demonstrations and source code are hosted at (https://candi.nextgenhealthcare.org), and (https://github.com/mbadge/candi), respectively, under GPL-3 license. SUPPLEMENTARY INFORMATION: Supplementary material is available at Bioinformatics online.
© The Author(s) 2018. Published by Oxford University Press.

Entities:  

Mesh:

Year:  2019        PMID: 30304439      PMCID: PMC6499410          DOI: 10.1093/bioinformatics/bty855

Source DB:  PubMed          Journal:  Bioinformatics        ISSN: 1367-4803            Impact factor:   6.931


1 Introduction

Computer vision algorithms have demonstrated success in many fields including medical radiology. Convolutional neural networks (CNNs) are a type of deep learning (DL) model that automatically learns image features and can be applied to several image-recognition tasks. Successful models are trained on the order of 100 000 training images acquired through multi-site efforts (Gulshan ; Ting ). In medicine, data collection and crowdsourcing are complicated by privacy and specialized training requirements. Web-based medical image annotation tools have been described but kept proprietary to an institution (Mata , 2017) and to specific crowdsourced projects (Cheplygina ; Maier-Hein ). LabelMe is a fully featured online tool designed for everyday images, but does not support sensitive data (Russell ). In clinical practice radiologists interpret images in the context of a patient’s previous image studies and non-image medical record data. There is a lack of annotation tools that provide multimodal patient data interfaces and can be deployed for collaborative work on sensitive data. Algorithms designed for Computer-Aided Diagnosis (CAD) are frequently only evaluated in isolation, and studies evaluating human performance with and without CAD have had inconsistent results. Retrospective studies on engineered feature (not DL) CAD in clinical practice have found accuracy benefit (Kasai ), no accuracy benefit (Benedikt ), or a negative effect (Gilbert ). CAD enhancement of human interpretation has been studied in disparate experimental designs. Commercially available CAD tools have been tested in fully randomized studies (Gilbert ) and observational studies (Fenton ). Experimental algorithms have been tested in only one mode (see RCT Case Study below) (Kasai ), or over multiple sessions (double-crossover design) where one day a radiologist interprets images with CAD and several months later she interprets images without CAD (or vice versa, by randomization) (Benedikt ). RCTs are graded as stronger evidence than pseudorandomized or observational studies, but RCTs have only been done with commercially available CAD systems. This manuscript introduces two open access computer-aided note and diagnosis interface (CANDI) web applications for collaboratively addressing the annotation and evaluation barriers to translating DL. The CANDI radiograph annotation dashboard (CANDI-RAD) app provides multimodal patient and image data to obtain training and testing data, and the CANDI-CAD evaluation app facilitates randomized controlled trials (RCTs) on human enhancement with algorithms.

2 Implementation

CANDI is distributed as an R package with web interfaces implemented as Shiny applications and modules which generate html and javascript browser-based dashboards. CANDI’s modules handle user input and render an image or all the images from a selected case, along with patient metadata. Additional modules for annotation graphically summarize a user’s entry records, and evaluation modules support CAD utilities (e.g. searching for similar images) and queue randomization. The package includes metadata from the public OpenI chest X-ray database (Demner-Fushman ) to demonstrate multimodal dashboards [images are separately available from the CC-NC-ND licensed openI database (https://openi.nlm.nih.gov/)]. We use third-party packages to support data input and output. The European Bioinformatics Institute package EBImage reads and renders standard biomedical image formats from disk or URL. The googlesheets package saves user input to the cloud for de-identified annotation storage. CANDI builds on these individual packages by providing Shiny modules so users can compile an interface suited for their study context (Badgeley ). Demonstration apps and user instructions are available at candi.nextgenhealthcare.org, which is hosted by a Nginx cloud server running Ubuntu. The CAD utilities were generated with several variations of Convolutional Neural Networks (CNNs) to predict disease status and localization and similar image search (further discussed in the Supplementary Material). The similar search module uses CNN image embeddings to compute the Euclidean distance between a test radiograph and all designated historical radiographs.

3 Case studies

3.1 Annotation

The CANDI training data generation app (candi.nextgenhealthcare.org/rad_institution) collects annotations for three supervised learning problems: (i) disease classification, (ii) image segmentation and (iii) image captioning. Each of these can be used to train a different implementation of a CNN (see Fig. 1). To adjudicate the gold-standard disease status, radiologists should use the multimodal app (candi.nextgenhealthcare.org/rad_case) to benefit from contemporaneous images and patients’ clinical data.
Fig. 1.

Annotation modalities and distinct uses. (A) The CANDI radiograph annotation (RAD) and computer-aided diagnosis (CAD) applications provide human-algorithm interfaces to generate training annotations and evaluate the subsequent models. Different annotation data modalities provide training data for distinct deep learning model utilities. We use convolutional neural networks (CNNs) to generate predictions in CANDI-CAD. (B) Various input/output systems are set up that conform to the security needs of different types of users

Annotation modalities and distinct uses. (A) The CANDI radiograph annotation (RAD) and computer-aided diagnosis (CAD) applications provide human-algorithm interfaces to generate training annotations and evaluate the subsequent models. Different annotation data modalities provide training data for distinct deep learning model utilities. We use convolutional neural networks (CNNs) to generate predictions in CANDI-CAD. (B) Various input/output systems are set up that conform to the security needs of different types of users

3.2 Evaluation randomized control trial

We implement CANDI-CAD to measure how users interpret radiographs under different assistance modes: concurrent and second-reader. In concurrent mode, the user receives algorithm support during the entire case interpretation, whereas in second-reader mode, algorithm support is only provided after the user formulates an initial unaided impression. Rigorous evaluation of new algorithms requires CAD software to be integrated into image database systems (Matsumoto ). CANDI-CAD enables experimental algorithms to be incorporated into image interpretation dashboards with randomized availability of CAD utilities. The demonstration at (candi.nextgenhealthcare.org/cad) uses three DL utilities: (i) Image similarity search, (ii) whole image classification and (iii) image bounding-box localization (see Fig. 1). Image queue order and CAD mode are fully randomized to facilitate a 2-arm RCT in one session.

4 Conclusion

CANDI aims to ease the translation of CAD algorithms to medical imaging by facilitating collaborative image annotation and randomized clinical evaluation. CANDI-RAD facilitates distributed annotation with a multimodal interface for patient context, which reflects clinical practice and allows radiologists to produce gold-standard data. CANDI-CAD facilitates randomized clinical trials to rigorously evaluate CAD augmentation of radiologists’ performance. Different data input/output interfaces can be used to apply CANDI to sensitive or public medical image data.

Funding

This work was supported by Verily Life Sciences, LLC as part of the Verily Academic Partnership with Icahn School of Medicine at Mount Sinai and by the National Institutes of Health, National Center for Advancing Translational Sciences (NCATS), Clinical and Translational Science Award [UL1TR001433-01] to J.T.D. Conflict of Interest: JTD has received consulting fees or honoraria from Janssen Pharmaceuticals, GlaxoSmithKline, AstraZeneca and Hoffman-La Roche. JTD is a scientific advisor to LAM Therapeutics and holds equity in NuMedii, Ayasdi and Ontomics. JL currently works for Merck in addition to his adjunct professor role at Boston University. Click here for additional data file.
  11 in total

1.  Crowdsourcing for reference correspondence generation in endoscopic images.

Authors:  Lena Maier-Hein; Sven Mersmann; Daniel Kondermann; Christian Stock; Hannes Gotz Kenngott; Alexandro Sanchez; Martin Wagner; Anas Preukschas; Anna-Laura Wekerle; Stefanie Helfert; Sebastian Bodenstedt; Stefanie Speidel
Journal:  Med Image Comput Comput Assist Interv       Date:  2014

2.  ProstateAnalyzer: Web-based medical application for the management of prostate cancer using multiparametric MR imaging.

Authors:  Christian Mata; Paul M Walker; Arnau Oliver; François Brunotte; Joan Martí; Alain Lalande
Journal:  Inform Health Soc Care       Date:  2015-02-24       Impact factor: 2.439

3.  Computer-aided detection of lung nodules on multidetector CT in concurrent-reader and second-reader modes: a comparative study.

Authors:  Sumiaki Matsumoto; Yoshiharu Ohno; Takatoshi Aoki; Hitoshi Yamagata; Munenobu Nogami; Keiko Matsumoto; Yoshiko Yamashita; Kazuro Sugimura
Journal:  Eur J Radiol       Date:  2013-03-06       Impact factor: 3.528

4.  Development and Validation of a Deep Learning Algorithm for Detection of Diabetic Retinopathy in Retinal Fundus Photographs.

Authors:  Varun Gulshan; Lily Peng; Marc Coram; Martin C Stumpe; Derek Wu; Arunachalam Narayanaswamy; Subhashini Venugopalan; Kasumi Widner; Tom Madams; Jorge Cuadros; Ramasamy Kim; Rajiv Raman; Philip C Nelson; Jessica L Mega; Dale R Webster
Journal:  JAMA       Date:  2016-12-13       Impact factor: 56.272

5.  Concurrent Computer-Aided Detection Improves Reading Time of Digital Breast Tomosynthesis and Maintains Interpretation Performance in a Multireader Multicase Study.

Authors:  Richard A Benedikt; Justin E Boatsman; Cynthia A Swann; Aaron D Kirkpatrick; Alicia Y Toledano
Journal:  AJR Am J Roentgenol       Date:  2017-10-24       Impact factor: 3.959

6.  Influence of computer-aided detection on performance of screening mammography.

Authors:  Joshua J Fenton; Stephen H Taplin; Patricia A Carney; Linn Abraham; Edward A Sickles; Carl D'Orsi; Eric A Berns; Gary Cutter; R Edward Hendrick; William E Barlow; Joann G Elmore
Journal:  N Engl J Med       Date:  2007-04-05       Impact factor: 91.245

7.  Preparing a collection of radiology examinations for distribution and retrieval.

Authors:  Dina Demner-Fushman; Marc D Kohli; Marc B Rosenman; Sonya E Shooshan; Laritza Rodriguez; Sameer Antani; George R Thoma; Clement J McDonald
Journal:  J Am Med Inform Assoc       Date:  2015-07-01       Impact factor: 4.497

8.  Usefulness of computer-aided diagnosis schemes for vertebral fractures and lung nodules on chest radiographs.

Authors:  Satoshi Kasai; Feng Li; Junji Shiraishi; Kunio Doi
Journal:  AJR Am J Roentgenol       Date:  2008-07       Impact factor: 3.959

9.  Single reading with computer-aided detection for screening mammography.

Authors:  Fiona J Gilbert; Susan M Astley; Maureen G C Gillan; Olorunsola F Agbaje; Matthew G Wallis; Jonathan James; Caroline R M Boggis; Stephen W Duffy
Journal:  N Engl J Med       Date:  2008-10-01       Impact factor: 91.245

10.  EHDViz: clinical dashboard development using open-source technologies.

Authors:  Marcus A Badgeley; Khader Shameer; Benjamin S Glicksberg; Max S Tomlinson; Matthew A Levin; Patrick J McCormick; Andrew Kasarskis; David L Reich; Joel T Dudley
Journal:  BMJ Open       Date:  2016-03-24       Impact factor: 2.692

View more
  2 in total

1.  High Precision Mammography Lesion Identification From Imprecise Medical Annotations.

Authors:  Ulzee An; Ankit Bhardwaj; Khader Shameer; Lakshminarayanan Subramanian
Journal:  Front Big Data       Date:  2021-12-03

2.  Fucose as a potential therapeutic molecule against the immune-mediated inflammation in IgA nepharopathy: An unrevealed link.

Authors:  Jianbo Qing; Xueli Hu; Changqun Li; Wenzhu Song; Hasna Tirichen; Hasnaa Yaigoub; Yafeng Li
Journal:  Front Immunol       Date:  2022-08-17       Impact factor: 8.786

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.