PURPOSE: Cancer stage is a key determinant of outcomes; however, stage is not available in claims-based data sources used for real-world evaluations. We compare multiple methods for classifying lung cancer stage from claims data. METHODS: Our study used the linked SEER-Medicare data. The patient samples included fee-for-service Medicare beneficiaries diagnosed with lung cancer from 2010 to 2011 (development cohort) and 2012 to 2013 (validation cohort) who received chemotherapy. Classification algorithms considered Medicare Part A and B claims for care in the 3 months before and after chemotherapy initiation. We developed a clinical algorithm to predict stage IV (v I to III) cancer on the basis of treatment patterns (surgery, radiotherapy, chemotherapy). We also considered an ensemble of claims-based machine learning algorithms. Classification methods were trained in the development cohort, and performance was measured in both cohorts. The SEER data were the gold standard for cancer stage. RESULTS: Development and validation cohorts included 14,760 and 14,620 patients with lung cancer, respectively. Validation analyses assessed clinical, random forest, and simple logistic regression algorithms. The best performing classifier within the development cohort was the random forests, but this performance was not replicated in validation analysis. Logistic regression had stable performance across cohorts. Compared with the clinical algorithm, the 14-variable logistic regression algorithm demonstrated higher accuracy in both the development (77% v 71%) and validation cohorts (77% v 73%), with improved specificity for stage IV disease. CONCLUSION: Machine learning algorithms have potential to improve lung cancer stage classification but may be prone to overfitting. Use of ensembles, cross-validation, and external validation can aid generalizability. Degradation of accuracy between development and validation cohorts suggests the need for caution in implementing machine learning in research or care delivery.
PURPOSE: Cancer stage is a key determinant of outcomes; however, stage is not available in claims-based data sources used for real-world evaluations. We compare multiple methods for classifying lung cancer stage from claims data. METHODS: Our study used the linked SEER-Medicare data. The patient samples included fee-for-service Medicare beneficiaries diagnosed with lung cancer from 2010 to 2011 (development cohort) and 2012 to 2013 (validation cohort) who received chemotherapy. Classification algorithms considered Medicare Part A and B claims for care in the 3 months before and after chemotherapy initiation. We developed a clinical algorithm to predict stage IV (v I to III) cancer on the basis of treatment patterns (surgery, radiotherapy, chemotherapy). We also considered an ensemble of claims-based machine learning algorithms. Classification methods were trained in the development cohort, and performance was measured in both cohorts. The SEER data were the gold standard for cancer stage. RESULTS: Development and validation cohorts included 14,760 and 14,620 patients with lung cancer, respectively. Validation analyses assessed clinical, random forest, and simple logistic regression algorithms. The best performing classifier within the development cohort was the random forests, but this performance was not replicated in validation analysis. Logistic regression had stable performance across cohorts. Compared with the clinical algorithm, the 14-variable logistic regression algorithm demonstrated higher accuracy in both the development (77% v 71%) and validation cohorts (77% v 73%), with improved specificity for stage IV disease. CONCLUSION: Machine learning algorithms have potential to improve lung cancer stage classification but may be prone to overfitting. Use of ensembles, cross-validation, and external validation can aid generalizability. Degradation of accuracy between development and validation cohorts suggests the need for caution in implementing machine learning in research or care delivery.
Authors: Neetu Chawla; K Robin Yabroff; Angela Mariotto; Timothy S McNeel; Deborah Schrag; Joan L Warren Journal: Ann Epidemiol Date: 2014-07-03 Impact factor: 3.797
Authors: Edward B Garon; Naiyer A Rizvi; Rina Hui; Natasha Leighl; Ani S Balmanoukian; Joseph Paul Eder; Amita Patnaik; Charu Aggarwal; Matthew Gubens; Leora Horn; Enric Carcereny; Myung-Ju Ahn; Enriqueta Felip; Jong-Seok Lee; Matthew D Hellmann; Omid Hamid; Jonathan W Goldman; Jean-Charles Soria; Marisa Dolled-Filhart; Ruth Z Rutledge; Jin Zhang; Jared K Lunceford; Reshma Rangwala; Gregory M Lubiniecki; Charlotte Roach; Kenneth Emancipator; Leena Gandhi Journal: N Engl J Med Date: 2015-04-19 Impact factor: 91.245
Authors: Beth L Nordstrom; Joanna L Whyte; Marilyn Stolar; Catherine Mercaldi; Joel D Kallich Journal: Pharmacoepidemiol Drug Saf Date: 2012-05 Impact factor: 2.890
Authors: Ronald M Kline; Carol Bazell; Erin Smith; Heidi Schumacher; Rahul Rajkumar; Patrick H Conway Journal: J Oncol Pract Date: 2015-02-17 Impact factor: 3.840
Authors: Scott J Antonia; Augusto Villegas; Davey Daniel; David Vicente; Shuji Murakami; Rina Hui; Takashi Yokoi; Alberto Chiappori; Ki H Lee; Maike de Wit; Byoung C Cho; Maryam Bourhaba; Xavier Quantin; Takaaki Tokito; Tarek Mekhail; David Planchard; Young-Chul Kim; Christos S Karapetis; Sandrine Hiret; Gyula Ostoros; Kaoru Kubota; Jhanelle E Gray; Luis Paz-Ares; Javier de Castro Carpeño; Catherine Wadsworth; Giovanni Melillo; Haiyi Jiang; Yifan Huang; Phillip A Dennis; Mustafa Özgüroğlu Journal: N Engl J Med Date: 2017-09-08 Impact factor: 91.245
Authors: Hava Izci; Tim Tambuyzer; Krizia Tuand; Victoria Depoorter; Annouschka Laenen; Hans Wildiers; Ignace Vergote; Liesbet Van Eycken; Harlinde De Schutter; Freija Verdoodt; Patrick Neven Journal: J Natl Cancer Inst Date: 2020-10-01 Impact factor: 13.506