| Literature DB >> 29992732 |
Hann-Hsiang Chao1, Gilmer Valdes1,2, Jose M Luna1, Marina Heskel1, Abigail T Berman1, Timothy D Solberg1,2, Charles B Simone3.
Abstract
BACKGROUND ANDEntities:
Keywords: zzm321990SBRTzzm321990; chest wall pain; dosimetry; machine learning; non-small-cell lung cancer
Mesh:
Year: 2018 PMID: 29992732 PMCID: PMC6123157 DOI: 10.1002/acm2.12415
Source DB: PubMed Journal: J Appl Clin Med Phys ISSN: 1526-9914 Impact factor: 2.102
List of all features selected for analysis. Each feature selected for analysis is listed and broken down by classification as a highly important or important feature
| Highly important features | Chest wall dose to 30 cc |
| Rib dose to 1 cc | |
| Rib dose max | |
| Medically inoperable vs patient refusal | |
| Dose per fraction | |
| Age | |
| Body mass index | |
| Tumor size (cm) | |
| PTV volume (cc) | |
| Age at first fraction | |
| Important features | Total dose |
| Diabetes (Y/N) | |
| Diffusion capacity of the lung for carbon monoxide (DLCO adj%) | |
| Forced expiratory volume (FEV1(L)) | |
| Decadron/prednisone use | |
| TNM status | |
| Stage | |
| Histology | |
| Lung mean dose | |
| Lung dose to 1000 cc | |
| Lung volume receiving 10 Gy | |
| Number of fractions |
Significant features identified on univariate analysis. Features with a CWS threshold with P < 0.05 (without adjustment for multiple comparisons) and generalization value > 0.75. The number of patients in each subgroup by feature threshold and the number and percentage of patients developing CWS in each subgroup are listed for reference. All features had missing values; therefore, the number of patients is <197 for each
| Feature | Thresholds | Subpopulations risks |
|
|---|---|---|---|
| Rib dose to 1 cc | <4000 cGy | ( | 0.010 |
| Chest wall dose 30 cc | <1900 cGy | ( | 0.035 |
| Lung dose to 1000 cc | <70 cGy | ( | 0.039 |
| Rib Dmax | <5100 cGy | ( | 0.050 |
Patient numbers do not add to 197 due to missing values present for select parameters.
Figure 1Modeling of optimal node size. The testing and training error for the dataset based on different hyperparameter settings of Min Number of observations per node using a leave‐one‐out cross‐validation of the deviance.
Figure 2Decision tree examples based on differential minimum node size. (a) Decision tree using a Min Number per Node = 50. (b) Decision tree using a Min Number per Node = 80.
Figure 3Multivariate analysis using hyperparameter adjustment and multiple iterations. Number of times features selected for 100 iterations and different hyperparameters. Min Number per Node = 50–80. In each iteration, 10–40 patients (in increments of 2) were randomly left out to introduce variation on the resulting decision tree. Features with a mean value equal to 0.1 of the feature that is selected the most are shown. The top of each bar represents the number of times each feature is selected per 100 iterations. Error bars represent 1 SD.
Figure 4Multivariate analysis using Random forests. Relative out‐of‐sample importance of different features calculating using Random forests. The top of the each bar represents the magnitude of out‐of‐sample importance for each feature. Error bars represent 1 SD.
Figure 5Learning curve for analysis dataset. Out‐of‐sample deviance estimated for training and testing data for different number of observations used for training. All points refer to the mean value over 100 iterations.