Literature DB >> 30229951

Comparative evaluation of autocontouring in clinical practice: A practical method using the Turing test.

Mark J Gooding1, Annamarie J Smith1, Maira Tariq1, Paul Aljabar1, Devis Peressutti1, Judith van der Stoep2, Bart Reymen2, Daisy Emans2, Djoya Hattu2, Judith van Loon2, Maud de Rooy2, Rinus Wanders2, Stephanie Peeters2, Tim Lustberg2, Johan van Soest2, Andre Dekker2, Wouter van Elmpt2.   

Abstract

PURPOSE: Automated techniques for estimating the contours of organs and structures in medical images have become more widespread and a variety of measures are available for assessing their quality. Quantitative measures of geometric agreement, for example, overlap with a gold-standard delineation, are popular but may not predict the level of clinical acceptance for the contouring method. Therefore, surrogate measures that relate more directly to the clinical judgment of contours, and to the way they are used in routine workflows, need to be developed. The purpose of this study is to propose a method (inspired by the Turing Test) for providing contour quality measures that directly draw upon practitioners' assessments of manual and automatic contours. This approach assumes that an inability to distinguish automatically produced contours from those of clinical experts would indicate that the contours are of sufficient quality for clinical use. In turn, it is anticipated that such contours would receive less manual editing prior to being accepted for clinical use. In this study, an initial assessment of this approach is performed with radiation oncologists and therapists.
METHODS: Eight clinical observers were presented with thoracic organ-at-risk contours through a web interface and were asked to determine if they were automatically generated or manually delineated. The accuracy of the visual determination was assessed, and the proportion of contours for which the source was misclassified recorded. Contours of six different organs in a clinical workflow were for 20 patient cases. The time required to edit autocontours to a clinically acceptable standard was also measured, as a gold standard of clinical utility. Established quantitative measures of autocontouring performance, such as Dice similarity coefficient with respect to the original clinical contour and the misclassification rate accessed with the proposed framework, were evaluated as surrogates of the editing time measured.
RESULTS: The misclassification rates for each organ were: esophagus 30.0%, heart 22.9%, left lung 51.2%, right lung 58.5%, mediastinum envelope 43.9%, and spinal cord 46.8%. The time savings resulting from editing the autocontours compared to the standard clinical workflow were 12%, 25%, 43%, 77%, 46%, and 50%, respectively, for these organs. The median Dice similarity coefficients between the clinical contours and the autocontours were 0.46, 0.90, 0.98, 0.98, 0.94, and 0.86, respectively, for these organs.
CONCLUSIONS: A better correspondence with time saving was observed for the misclassification rate than the quantitative contour measures explored. From this, we conclude that the inability to accurately judge the source of a contour indicates a reduced need for editing and therefore a greater time saving overall. Hence, task-based assessments of contouring performance may be considered as an additional way of evaluating the clinical utility of autosegmentation methods.
© 2018 American Association of Physicists in Medicine.

Entities:  

Keywords:  Turing test; assessment; autocontouring; editing time; organs-at-risk

Mesh:

Year:  2018        PMID: 30229951     DOI: 10.1002/mp.13200

Source DB:  PubMed          Journal:  Med Phys        ISSN: 0094-2405            Impact factor:   4.071


  20 in total

1.  Auto-detection and segmentation of involved lymph nodes in HPV-associated oropharyngeal cancer using a convolutional deep learning neural network.

Authors:  Nicolette Taku; Kareem A Wahid; Lisanne V van Dijk; Jaakko Sahlsten; Joel Jaskari; Kimmo Kaski; Clifton D Fuller; Mohamed A Naser
Journal:  Clin Transl Radiat Oncol       Date:  2022-06-18

2.  Patient specific deep learning based segmentation for magnetic resonance guided prostate radiotherapy.

Authors:  Samuel Fransson; David Tilly; Robin Strand
Journal:  Phys Imaging Radiat Oncol       Date:  2022-06-03

3.  External validation of deep learning-based contouring of head and neck organs at risk.

Authors:  Ellen J L Brunenberg; Isabell K Steinseifer; Sven van den Bosch; Johannes H A M Kaanders; Charlotte L Brouwer; Mark J Gooding; Wouter van Elmpt; René Monshouwer
Journal:  Phys Imaging Radiat Oncol       Date:  2020-07-10

Review 4.  Metrics to evaluate the performance of auto-segmentation for radiation treatment planning: A critical review.

Authors:  Michael V Sherer; Diana Lin; Sharif Elguindi; Simon Duke; Li-Tee Tan; Jon Cacicedo; Max Dahele; Erin F Gillespie
Journal:  Radiother Oncol       Date:  2021-05-11       Impact factor: 6.901

5.  Fully Automated 3-D Ultrasound Segmentation of the Placenta, Amniotic Fluid, and Fetus for Early Pregnancy Assessment.

Authors:  Padraig Looney; Yi Yin; Sally L Collins; Kypros H Nicolaides; Walter Plasencia; Malid Molloholli; Stavros Natsis; Gordon N Stevenson
Journal:  IEEE Trans Ultrason Ferroelectr Freq Control       Date:  2021-05-25       Impact factor: 3.267

6.  Clinical implementation of deep learning contour autosegmentation for prostate radiotherapy.

Authors:  Elaine Cha; Sharif Elguindi; Ifeanyirochukwu Onochie; Daniel Gorovets; Joseph O Deasy; Michael Zelefsky; Erin F Gillespie
Journal:  Radiother Oncol       Date:  2021-03-03       Impact factor: 6.901

7.  Considerations for artificial intelligence clinical impact in oncologic imaging: an AI4HI position paper.

Authors:  Luis Marti-Bonmati; Dow-Mu Koh; Katrine Riklund; Maciej Bobowicz; Yiannis Roussakis; Joan C Vilanova; Jurgen J Fütterer; Jordi Rimola; Pedro Mallol; Gloria Ribas; Ana Miguel; Manolis Tsiknakis; Karim Lekadir; Gianna Tsakou
Journal:  Insights Imaging       Date:  2022-05-10

Review 8.  Applications and limitations of machine learning in radiation oncology.

Authors:  Daniel Jarrett; Eleanor Stride; Katherine Vallis; Mark J Gooding
Journal:  Br J Radiol       Date:  2019-06-05       Impact factor: 3.629

9.  Evaluation of measures for assessing time-saving of automatic organ-at-risk segmentation in radiotherapy.

Authors:  Femke Vaassen; Colien Hazelaar; Ana Vaniqui; Mark Gooding; Brent van der Heyden; Richard Canters; Wouter van Elmpt
Journal:  Phys Imaging Radiat Oncol       Date:  2019-12-17

Review 10.  Artificial intelligence-based clinical decision support in modern medical physics: Selection, acceptance, commissioning, and quality assurance.

Authors:  Geetha Mahadevaiah; Prasad Rv; Inigo Bermejo; David Jaffray; Andre Dekker; Leonard Wee
Journal:  Med Phys       Date:  2020-06       Impact factor: 4.071

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.