| Literature DB >> 30175071 |
Hesham Elhalawani1, Timothy A Lin1,2, Stefania Volpe1,3, Abdallah S R Mohamed1,4, Aubrey L White1,5, James Zafereo1,5, Andrew J Wong1,6, Joel E Berends1,6, Shady AboHashem1,7, Bowman Williams1,8, Jeremy M Aymard1,9, Aasheesh Kanwar1,10, Subha Perni1,11, Crosby D Rock1,12, Luke Cooksey1,13, Shauna Campbell1,14, Pei Yang1,2, Khahn Nguyen15, Rachel B Ger16,17, Carlos E Cardenas16,17, Xenia J Fave18, Carlo Sansone19, Gabriele Piantadosi19, Stefano Marrone19, Rongjie Liu2,20, Chao Huang2,20, Kaixian Yu2,20, Tengfei Li2,20, Yang Yu2,20, Youyi Zhang2,20, Hongtu Zhu2,20, Jeffrey S Morris2,20, Veerabhadran Baladandayuthapani2,20, John W Shumway1, Alakonanda Ghosh1, Andrei Pöhlmann21, Hady A Phoulady22, Vibhas Goyal23, Guadalupe Canahuate24, G Elisabeta Marai25, David Vock26, Stephen Y Lai27, Dennis S Mackin15,17, Laurence E Court15,17, John Freymann28, Keyvan Farahani29,30, Jayashree Kaplathy-Cramer31, Clifton D Fuller1,2,17.
Abstract
Radiomics leverages existing image datasets to provide non-visible data extraction via image post-processing, with the aim of identifying prognostic, and predictive imaging features at a sub-region of interest level. However, the application of radiomics is hampered by several challenges such as lack of image acquisition/analysis method standardization, impeding generalizability. As of yet, radiomics remains intriguing, but not clinically validated. We aimed to test the feasibility of a non-custom-constructed platform for disseminating existing large, standardized databases across institutions for promoting radiomics studies. Hence, University of Texas MD Anderson Cancer Center organized two public radiomics challenges in head and neck radiation oncology domain. This was done in conjunction with MICCAI 2016 satellite symposium using Kaggle-in-Class, a machine-learning and predictive analytics platform. We drew on clinical data matched to radiomics data derived from diagnostic contrast-enhanced computed tomography (CECT) images in a dataset of 315 patients with oropharyngeal cancer. Contestants were tasked to develop models for (i) classifying patients according to their human papillomavirus status, or (ii) predicting local tumor recurrence, following radiotherapy. Data were split into training, and test sets. Seventeen teams from various professional domains participated in one or both of the challenges. This review paper was based on the contestants' feedback; provided by 8 contestants only (47%). Six contestants (75%) incorporated extracted radiomics features into their predictive model building, either alone (n = 5; 62.5%), as was the case with the winner of the "HPV" challenge, or in conjunction with matched clinical attributes (n = 2; 25%). Only 23% of contestants, notably, including the winner of the "local recurrence" challenge, built their model relying solely on clinical data. In addition to the value of the integration of machine learning into clinical decision-making, our experience sheds light on challenges in sharing and directing existing datasets toward clinical applications of radiomics, including hyper-dimensionality of the clinical/imaging data attributes. Our experience may help guide researchers to create a framework for sharing and reuse of already published data that we believe will ultimately accelerate the pace of clinical applications of radiomics; both in challenge or clinical settings.Entities:
Keywords: big data; head and neck; machine learning; radiation oncology; radiomics challenge
Year: 2018 PMID: 30175071 PMCID: PMC6107800 DOI: 10.3389/fonc.2018.00294
Source DB: PubMed Journal: Front Oncol ISSN: 2234-943X Impact factor: 6.244
Supplemental information about data provided for radiomics challenges.
| Patient ID | Numbers given randomly to the patient after anonymization of the DICOM protected health identifier (PHI) tag (0010,0020) that corresponds to medical record number |
| HPV/p16 status | HPV status, as assessed by HPV DNA |
| Gender | Patient's sex |
| Age at diagnosis | Patient's age in years at the time of diagnosis |
| Race | American Indian/Alaska Native, Asian, Black, Hispanic, White, or not applicable |
| Tumor laterality | Right, left, or bilateral |
| Oropharynx subsite of origin | Subsite of the tumor within the oropharynx, i.e., base of tongue (21) or tonsil/soft palate/pharyngeal wall/glossopharyngeal sulcus/other (no single subsite of origin could be identified) |
| T category | Description of the original (primary) tumor with regard to size and extent per the American Joint Committee on Cancer (AJCC) and Union for International Cancer Control (UICC) cancer staging system, i.e., T1, T2, T3, or T4 ( |
| N category | Description of whether the cancer has reached nearby lymph nodes, per the AJCC and UICC cancer staging system, i.e., N0, N1, N2a, N2b, N2c, or N3 ( |
| AJCC stage | AJCC cancer stage ( |
| Pathologic grade | Grade of tumor differentiation, i.e., I, II, III, IV, I-II, II-III, or not assessable |
| Smoking status at diagnosis | Never, current, or former smoker |
| Smoking pack-years | An equivalent numerical value of lifetime tobacco exposure; 1 pack-year is defined as 20 cigarettes smoked every day for 1 year |
Figure 1Workflow of radiomics challenges.
Figure 2Diagram illustrating the splitting of datasets per the challenge's rules.
Challenges and derived lessons from organizing open-source radiomics challenges.
| • Paucity of open-source freely available radiomics datasets |
| • Use common ontology guidelines to assign nomenclature for target volumes and clinical data |
| • Adopt “Public/Private leaderboard” challenges to mitigate overtraining/overfitting |