Literature DB >> 31962297

Data clustering to select clinically-relevant test cases for algorithm benchmarking and characterization.

Sarah Weppler1, Colleen Schinkel, Charles Kirkby, Wendy Smith.   

Abstract

Algorithm benchmarking and characterization are an important part of algorithm development and validation prior to clinical implementation. However, benchmarking may be limited to a small collection of test cases due to the resource-intensive nature of establishing 'ground-truth' references. This study proposes a framework for selecting test cases to assess algorithm and workflow equivalence. Effective test case selection may minimize the number of ground-truth comparisons required to establish robust and clinically relevant benchmarking and characterization results. To demonstrate the proposed framework, we clustered differences between two independent workflows estimating during-treatment dose objective violations for 15 head and neck cancer patients (15 planning CTs, 105 on-unit CBCTs). Each workflow used a different deformable image registration algorithm to estimate inter-fractional anatomy and contour changes. The Hopkins statistic tested whether workflow output was inherently clustered and k-medoid clustering formalized cluster assignment. Further statistical analyses verified the relevance of clusters to algorithm output. Data at cluster centers ('medoids') were considered as candidate test cases representative of workflow-relevant algorithm differences. The framework indicated that differences in estimated dose objective violations were naturally grouped (Hopkins  =  0.75, providing 90% confidence). K-medoid clustering identified five clusters which stratified workflow differences (MANOVA: p   <  0.001) in estimated parotid gland D50%, spinal cord/brainstem Dmax, and high dose CTV coverage dose violations (Kendall's tau: p   <  0.05). Systematic algorithm differences resulting in workflow discrepancies were: parotid gland volumes (ANOVA: p   <  0.001), external contour deformations (t-test: p   =  0.022), and CTV-to-PTV margins (t-test: 0.009), respectively. Five candidate test cases were verified as representative of the five clusters. The framework successfully clustered workflow outputs and identified five test cases representative of clinically relevant algorithm discrepancies. This approach may improve the allocation of resources during the benchmarking and characterization process and the applicability of results to clinical data.

Entities:  

Year:  2020        PMID: 31962297     DOI: 10.1088/1361-6560/ab6e54

Source DB:  PubMed          Journal:  Phys Med Biol        ISSN: 0031-9155            Impact factor:   3.609


  2 in total

1.  Patient-Reported Outcomes-Guided Adaptive Radiation Therapy for Head and Neck Cancer.

Authors:  Sarah Weppler; Harvey Quon; Colleen Schinkel; Adam Yarschenko; Lisa Barbera; Nabhya Harjai; Wendy Smith
Journal:  Front Oncol       Date:  2021-10-19       Impact factor: 6.244

2.  Determining Clinical Patient Selection Guidelines for Head and Neck Adaptive Radiation Therapy Using Random Forest Modelling and a Novel Simplification Heuristic.

Authors:  Sarah Weppler; Harvey Quon; Colleen Schinkel; James Ddamba; Nabhya Harjai; Clarisse Vigal; Craig A Beers; Lukas Van Dyke; Wendy Smith
Journal:  Front Oncol       Date:  2021-06-07       Impact factor: 6.244

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.