| Literature DB >> 36160349 |
Cheng Xu1,2, Jing Wang1, Tianlong Zheng1, Yue Cao3, Fan Ye4.
Abstract
Introduction: It is essential to predict the survival status of patients based on their prognosis. This can assist physicians in evaluating treatment decisions. Random forest is an excellent machine learning algorithm even without any modification. We propose a new random forest weighting method and apply it to the gastric cancer patient data from the Surveillance, Epidemiology, and End Results (SEER) program. We evaluated the generalization ability of this weighted random forest algorithm on 10 public medical datasets. Furthermore, for the same weighting mode, the difference between using out-of-bag (OOB) data and all training sets as the weighting basis is explored. Material and methods: 110 697 cases of gastric cancer patients diagnosed between 1975 and 2016 obtained from the SEER database were included in the experiment. In addition, 10 public medical datasets were used for the generalization ability evaluation of this weighted random forest algorithm.Entities:
Keywords: SEER; algorithm improvement; gastric cancer; machine learning; prognosis; random forest; survival time
Year: 2021 PMID: 36160349 PMCID: PMC9479734 DOI: 10.5114/aoms/135594
Source DB: PubMed Journal: Arch Med Sci ISSN: 1734-1922 Impact factor: 3.707
Weighting methods for ensemble classifiers in the literature
| Work | Method applied | Conclusion |
|---|---|---|
| [ | Tree-level weights in random forest | This method cannot significantly improve the predictive ability of high-dimensional genetic data, but it can improve performance in other fields |
| [ | Variable importance-weighted random forest | Improved accuracy compared to the original random forest |
| [ | Refined weighted random forest | All data (in-bag data and out-of-bag data) were used for training, and more accurate than the original random forest |
| [ | Exponentially weighted random forest | Calculate the similarity between each test example and the decision tree from the random forest to use as a weight. The results show that it is better than all random forests on most data sets |
| [ | Stacking-based random forest models | Four weighted improvement methods are proposed, they are based on k-fold cross-validation, AUC value of a single tree, OOB data measurement accuracy, and stacking-based random forest models |
| [ | Cesáro averages for weighted trees | On the decision tree, replace the regular average with a Cesáro average. There is an improvement between 0.2% and 0.5% on the data set listed by the author |
| [ | Weight assignment based on error rate of OOB data | In this approach, the authors assign the weight of each tree according to the relationship between the error rate of OOB data of each tree and the average OOB data error rate |
| [ | Weighted vote for trees aggregation | Each tree is evaluated by OOB data and used as a weight, and the classification result is obtained by comparing the aggregation of weight of each class |
Figure 1Random forest model
Figure 2OOBWRF model
Figure 3TLWRF model
Figure 4Comparison of work details between RF and TLWRF
Details of the results of data characteristics analysis of gastric cancer patients based on Cox regression model (n = 110 697)
| Variables | Coef | Exp(coef) | Se(coef) | Coef lower 95% | Coef upper 95% | Exp(coef) lower 95% | Exp(coef) upper 95% |
| Log2(p) | |
|---|---|---|---|---|---|---|---|---|---|---|
| Age at diagnosis | 0.01 | 1.01 | 0.00 | 0.01 | 0.01 | 1.01 | 1.01 | 40.58 | < 0.005 | Inf |
| Sex | 0.07 | 1.07 | 0.01 | 0.05 | 0.08 | 1.05 | 1.08 | 9.50 | < 0.005 | 68.75 |
| Race recode (W, B, AI, API) | –0.05 | 0.95 | 0.00 | –0.06 | –0.04 | 0.95 | 0.96 | –11.48 | < 0.005 | 98.94 |
| Marital status at diagnosis | 0.06 | 1.06 | 0.00 | 0.05 | 0.06 | 1.05 | 1.07 | 16.77 | < 0.005 | 207.22 |
| Insurance Recode (2007+) | –0.02 | 0.98 | 0.00 | –0.02 | –0.02 | 0.98 | 0.98 | –12.58 | < 0.005 | 118.18 |
| PRCDA 2016 | 0.01 | 1.01 | 0.01 | 0.00 | 0.03 | 1.00 | 1.03 | 1.48 | 0.14 | 2.85 |
| Primary site | 0.02 | 1.02 | 0.00 | 0.02 | 0.02 | 1.02 | 1.02 | 17.06 | < 0.005 | 214.34 |
| Grade | 0.06 | 1.06 | 0.00 | 0.05 | 0.07 | 1.06 | 1.07 | 20.61 | < 0.005 | 310.97 |
| CS tumor size (2004–2015) | 0.00 | 1.00 | 0.00 | 0.00 | 0.00 | 1.00 | 1.00 | 23.44 | < 0.005 | 401.08 |
| CS extension (2004–2015) | 0.00 | 1.00 | 0.00 | 0.00 | 0.00 | 1.00 | 1.00 | 10.70 | < 0.005 | 86.35 |
| CS lymph nodes (2004–2015) | 0.00 | 1.00 | 0.00 | 0.00 | 0.00 | 1.00 | 1.00 | 8.22 | < 0.005 | 52.09 |
| CS mets at dx (2004–2015) | 0.00 | 1.00 | 0.00 | 0.00 | 0.00 | 1.00 | 1.00 | 3.85 | < 0.005 | 13.07 |
| COD to site recode | –0.01 | 0.99 | 0.00 | –0.01 | –0.01 | 0.99 | 0.99 | –47.41 | < 0.005 | inf |
| Sequence number | –0.01 | 0.99 | 0.02 | –0.05 | 0.02 | 0.96 | 1.02 | –0.64 | 0.52 | 0.93 |
| First malignant primary indicator | –0.15 | 0.86 | 0.02 | –0.19 | –0.10 | 0.82 | 0.91 | –5.88 | < 0.005 | 27.87 |
| Total number of in situ/malignant tumors for patient | –0.13 | 0.88 | 0.02 | –0.16 | –0.09 | 0.85 | 0.91 | –7.31 | < 0.005 | 41.77 |
| Total number of benign/borderline tumors for patient | 0.02 | 1.02 | 0.09 | –0.15 | 0.19 | 0.86 | 1.20 | 0.20 | 0.84 | 0.24 |
| Derived AJCC T, 7th ed (2010–2015) | –0.07 | 0.94 | 0.01 | –0.08 | –0.05 | 0.93 | 0.95 | –10.66 | < 0.005 | 85.69 |
| Derived AJCC N, 7th ed (2010–2015) | –0.12 | 0.89 | 0.01 | –0.13 | –0.10 | 0.88 | 0.90 | –16.19 | < 0.005 | 193.42 |
| Derived AJCC M, 7th ed (2010–2015) | –0.11 | 0.89 | 0.02 | –0.15 | –0.08 | 0.86 | 0.93 | –6.09 | < 0.005 | 29.74 |
| Derived AJCC Stage Group, 7th ed (2010–2015) | 0.15 | 1.16 | 0.00 | 0.14 | 0.16 | 1.15 | 1.17 | 35.47 | < 0.005 | 912.78 |
Details of target attribute distribution
| Variable | 3 Years | 5 Years | 10 Years | >10 Years |
|---|---|---|---|---|
| Original | 82247 (85.99%) | 4836 (5.06%) | 4754 (4.97%) | 3811 (3.98%) |
| Oversampled | 82247 (25.00%) | 82247 (25.00%) | 82247 (25.00%) | 82247 (25.00%) |
Experimental I results of OOBWRF and TLWRF compared to DT and original RF. The best-performing classifier for each evaluation index is highlighted
| Variable | DT | RF | OOBWRF | TLWRF |
|---|---|---|---|---|
| Accuracy | 79.81% | 85.12% | 85.55% | 85.91% |
| Class 0 AUC | 64.23% | 70.29% | 72.92% | 80.88% |
| Class 1 AUC | 52.14% | 67.63% | 67.88% | 65.55% |
| Class 2 AUC | 55.99% | 79.00% | 79.37% | 77.86% |
| Class 3 AUC | 66.00% | 89.52% | 90.60% | 91.45% |
| Macro-average AUC | 59.59% | 76.61% | 77.70% | 78.94% |
| Micro-average AUC | 86.31% | 95.18% | 95.38% | 95.69% |
Figure 5Details of the ROC curve of each model
Details of the test datasets downloaded from the UCI machine learning repository
| Number | Name | Size | Classes | Features |
|---|---|---|---|---|
| 1 | Adult | 32561 | 2 | 14 |
| 2 | Breast Cancer (Coimbra) | 116 | 2 | 9 |
| 3 | Breast Cancer (Michalski) | 286 | 2 | 9 |
| 4 | Breast Cancer (Wisconsin) | 683 | 2 | 10 |
| 5 | Extension of Z-Alizadeh sani | 303 | 2 | 55 |
| 6 | Haberman’s Survival Data Set | 306 | 2 | 3 |
| 7 | Heart failure clinical records Data | 583 | 2 | 10 |
| 8 | Dermatology | 366 | 6 | 34 |
| 9 | Ecoli | 336 | 8 | 8 |
| 10 | Lymphography Data Set | 148 | 4 | 18 |
Experimental II results of OOBWRF and TLWRF compared to DT and original RF. The best-performing classifier for each dataset is highlighted according to accuracy and AUC. The last row shows the average accuracy of all models considering all datasets
| Number | Accuracy | AUC | ||||||
|---|---|---|---|---|---|---|---|---|
| DT | RF | OOBWRF | TLWRF | DT | RF | OOBWRF | TLWRF | |
| 1 | 81.09% | 85.34% | 84.53% | 85.54% | 74.86% | 90.60% | 88.82% | 89.70% |
| 2 | 62.86% | 62.86% | 63.35% | 65.71% | 62.42% | 74.84% | 74.93% | 75.18% |
| 3 | 70.24% | 78.57% | 80.95% | 80.95% | 57.26% | 72.47% | 69.90% | 71.83% |
| 4 | 93.66% | 95.12% | 92.68% | 95.12% | 92.46% | 99.37% | 99.37% | 99.36% |
| 5 | 72.53% | 81.32% | 74.73% | 82.42% | 68.69% | 92.54% | 92.16% | 92.59% |
| 6 | 60.87% | 63.04% | 65.22% | 65.22% | 62.75% | 73.91% | 62.75% | 79.45% |
| 7 | 66.29% | 68.57% | 70.29% | 69.71% | 59.28% | 72.19% | 72.19% | 73.29% |
| 8 | 87.96% | 95.37% | 89.81% | 96.30% | 95.71% | 99.80% | 98.95% | 99.40% |
| 9 | 78.22% | 85.15% | 72.28% | 86.55% | 76.07% | 84.46% | 86.06% | 90.56% |
| 10 | 75.56% | 84.44% | 84.44% | 86.67% | 79.91% | 97.62% | 97.62% | 98.44% |
| Mean | 74.93% | 79.98% | 77.83% | 81.42% | 72.94% | 85.78% | 84.27% | 86.98% |