Literature DB >> 31140124

Generating Real-World Tumor Burden Endpoints from Electronic Health Record Data: Comparison of RECIST, Radiology-Anchored, and Clinician-Anchored Approaches for Abstracting Real-World Progression in Non-Small Cell Lung Cancer.

Sandra D Griffith1, Melisa Tucker1, Bryan Bowser1, Geoffrey Calkins1, Che-Hsu Joe Chang1, Ellie Guardino2, Sean Khozin3, Josh Kraut1, Paul You1, Deb Schrag4,5, Rebecca A Miksad6.   

Abstract

INTRODUCTION: Real-world evidence derived from electronic health records (EHRs) is increasingly recognized as a supplement to evidence generated from traditional clinical trials. In oncology, tumor-based Response Evaluation Criteria in Solid Tumors (RECIST) endpoints are standard clinical trial metrics. The best approach for collecting similar endpoints from EHRs remains unknown. We evaluated the feasibility of a RECIST-based methodology to assess EHR-derived real-world progression (rwP) and explored non-RECIST-based approaches.
METHODS: In this retrospective study, cohorts were randomly selected from Flatiron Health's database of de-identified patient-level EHR data in advanced non-small cell lung cancer. A RECIST-based approach tested for feasibility (N = 26). Three non-RECIST approaches were tested for feasibility, reliability, and validity (N = 200): (1) radiology-anchored, (2) clinician-anchored, and (3) combined. Qualitative and quantitative methods were used.
RESULTS: A RECIST-based approach was not feasible: cancer progression could be ascertained for 23% (6/26 patients). Radiology- and clinician-anchored approaches identified at least one rwP event for 87% (173/200 patients). rwP dates matched 90% of the time. In 72% of patients (124/173), the first clinician-anchored rwP event was accompanied by a downstream event (e.g., treatment change); the association was slightly lower for the radiology-anchored approach (67%; 121/180). Median overall survival (OS) was 17 months [95% confidence interval (CI) 14, 19]. Median real-world progression-free survival (rwPFS) was 5.5 months (95% CI 4.6, 6.3) and 4.9 months (95% CI 4.2, 5.6) for clinician-anchored and radiology-anchored approaches, respectively. Correlations between rwPFS and OS were similar across approaches (Spearman's rho 0.65-0.66). Abstractors preferred the clinician-anchored approach as it provided more comprehensive context.
CONCLUSIONS: RECIST cannot adequately assess cancer progression in EHR-derived data because of missing data and lack of clarity in radiology reports. We found a clinician-anchored approach supported by radiology report data to be the optimal, and most practical, method for characterizing tumor-based endpoints from EHR-sourced data. FUNDING: Flatiron Health Inc., which is an independent subsidiary of the Roche group.

Entities:  

Keywords:  Carcinoma, non-small cell lung; Endpoints; Immunotherapy; PD-1; PD-L1; Real-world evidence

Mesh:

Year:  2019        PMID: 31140124      PMCID: PMC6822856          DOI: 10.1007/s12325-019-00970-1

Source DB:  PubMed          Journal:  Adv Ther        ISSN: 0741-238X            Impact factor:   3.845


Introduction

As oncology clinical trials have become increasingly complex, their efficacy results may not directly translate into real-world effectiveness [1]. Real-world evidence (RWE) sources, including data derived from electronic health records (EHRs), can augment insights from traditional clinical trials by better understanding the realities of real-world patients [2, 3]. RWE can be leveraged to improve therapy development programs, increasing the external validity of evidence available to support treatment decisions, and the FDA recently issued the Framework for its Real-World Evidence Program [4-7]. The key to fully unlocking the value of the EHR for these purposes is reliable and obtainable outcome metrics for patients in real-world settings. However, whether traditional cancer endpoints other than mortality can be gleaned from EHR data remains unknown. We need to determine whether we can apply traditionally defined clinical trial tumor endpoints to EHR data, or whether new definitions are needed. Endpoints based on tumor size changes are often treatment efficacy metrics in solid tumor clinical trials [8, 9], where radiographic images are evaluated using the Response Evaluation Criteria in Solid Tumors (RECIST) and reviewed by an independent central committee to improve assessment objectivity [10, 11]. RECIST assesses size changes of designated tumors over time (“target lesions”) combined with the presence or absence of new tumors to categorize patients’ disease status: “response,” “stable disease,” or “progression.” Routine clinical practice follows a similar assessment paradigm. Treatment effectiveness is often determined by periodically evaluating a variety of clinical parameters (e.g., imaging, physical exam, biomarkers, pathology specimen, patient-reported concerns). Depending on the clinical context, EHR documentation of changes in one or more of these parameters may reflect the outcome of an intervention and tumor burden dynamics may be more globally summarized as “improved” (i.e., “tumor response”), “no change” (i.e., “stable disease”), or “worse” (i.e., “tumor progression,” an event for which the patient may need a new treatment). In addition, outcome appraisals recorded in the EHR may also incorporate quality of life metrics, and other quantitative and qualitative assessments and test results. EHR data provide significant opportunities to develop real-world data (RWD)-specific metrics but may also add challenges to the interpretation of real-world outcomes. For example, the relevant information may reside within structured sources (i.e., highly organized data, such as white blood cell count) or unstructured (e.g., free-text clinical notes or radiology or biomarker reports). In other words, the EHR holds information, but differently from a clinical trial data set. Therefore, we may need to process the available EHR data in a different way, culling it through manual curation by trained abstractor experts following precise, pre-defined policies and procedures, through computer-based algorithms mimicking the manual approach or through a combination of both. Which EHR evidence source is best to anchor cancer outcomes against? Radiology reports may not include declarative RECIST assessments but typically have an “Impression” section written by the reviewing radiologist. Clinician assessments, findings, and interpretation of relevant results and reports are often summarized in EHR notes by the clinician caring for the patient, especially when they inform treatment decisions. Therefore, clinician notes may serve as an alternative or supplement to radiology reports to curate cancer outcomes from EHR data. Our objective was to identify a practical and efficient large-scale data abstraction method for estimation of cancer outcomes from the EHR. As a result of the variety of EHR formats and locations where documentation of outcome assessments may exist, such an abstraction approach must (1) be applicable across multiple EHR systems; (2) be amenable to manual and/or electronic abstraction from unstructured documents; (3) accommodate clinical judgement nuances; (4) be reliable despite possibly missing data points; and (5) be efficient to support scaling for large cohort assessments. We anticipated that traditional clinical trial approaches to collecting endpoints, such as cancer progression, may need to be modified for the unique features of EHR data. We tested this hypothesis in a preliminary experiment in a small cohort of patients with advanced non-small cell lung cancer (aNSCLC). We then compared several alternative approaches for abstraction of cancer progression events from the EHR in a larger cohort of aNSCLC patients.

Methods

We conducted experiments to answer the following questions: (1) can RECIST be feasibly applied to EHR cancer progression data, and (2) how do alternate abstraction approaches perform?

Data Sources

The overall cohort was selected from Flatiron Health’s longitudinal EHR-derived database, which, at the time of this study, included over 210 cancer clinics representing more than 1.2 million active patients across the USA. We identified more than 120,000 and 25,000 patients diagnosed with lung cancer and aNSCLC, respectively. The majority of patients were treated in community oncology clinics. Demographic, clinical, and outcomes data were extracted out of the source EHR, including structured data and unstructured documents. To create the database, we aggregated, normalized, de-identified, and harmonized patient-level data. Data were processed centrally and stored in a secure format and environment. Structured data (e.g., treatments, labs, diagnosis codes) were mapped to standard ontologies. Dates of death were obtained from a composite mortality variable comprising the EHR structured data linked to commercial mortality data and the Social Security Death Index [12]. Unstructured data (e.g., clinician notes, radiology reports, death notices) were extracted from EHR-based digital documents via “technology-enabled” chart abstraction [13]. Every data point sourced from unstructured documents was manually reviewed by trained chart abstractors (clinical oncology nurses and tumor registrars, with oversight from medical oncologists). Quality control included duplicate chart abstraction of a sample of abstracted variables as well as logic checks based on clinical and data considerations.

Study Design and Data Collection

This retrospective observational study investigated methods to assess cancer progression from EHR data through two objectives (Fig. 1): (1) to evaluate the feasibility of a RECIST approach (experiment 1) and (2) if a RECIST approach is not feasible, to evaluate three alternative non-RECIST abstraction approaches (experiment 2).
Fig. 1

Using the EHR to generate a cancer progression endpoint

Using the EHR to generate a cancer progression endpoint Inclusion criteria for the study cohort from which patients were randomly selected for both experiments were (1) NSCLC patients diagnosed with advanced disease between January 1, 2011 and April 1, 2016; (2) at least two clinical visits on or after January 1, 2011 documented in the EHR; and (3) documentation of initiation of at least two lines of systemic therapy after advanced diagnosis. Advanced NSCLC was defined as a diagnosis of stage IIIB or metastatic stage IV, or recurrent early disease. At the time of this study, staging of patients diagnosed with aNSCLC followed the criteria from the American Joint Commission of Cancer/Union for International Cancer Control/International Association for the Study of Lung Cancer (AJCC/UICC/IASLC) 7th edition manual [14]. To assess the feasibility of the RECIST approach (experiment 1), a total of 26 patients were randomly selected. The sample size was chosen to achieve expected theme saturation consistent with an initial feasibility study to assess usability [15, 16]. To study the non-RECIST abstraction approaches (experiment 2), 200 patients were randomly selected. The sample size was chosen to balance feasibility and the need for a sufficient number of progression and death events to inform descriptive analyses. Institutional review board (IRB) approval of the study protocol (IRB # RWE-001, “The Flatiron Health Real-World Evidence Parent Protocol”, Tracking # FLI1-18-044) by the Copernicus Group IRB, with waiver of informed consent, was obtained prior to study conduct, and covers the data from all sites represented.

Experiment 1: Feasibility of RECIST Criteria

To determine if the elements required for RECIST version 1.1 [11] evaluation could be abstracted from patient charts (i.e., usability), we evaluated radiology reports in the EHR for baseline and first treatment response imaging assessment after first line (1L) systemic therapy initiation. Radiology reports were evaluated for explicit descriptions of “target lesions” as required by RECIST. The subgroup of patients for whom an explicit description of target lesions was not found were also evaluated with a more lenient approach: abstractors identified target lesions on the basis of available radiology reports. Lesions of at least 1 cm were classified as measured if size was numerically documented in the radiology report. The RECIST approach was determined to be potentially feasible for a patient if (1) their baseline scan was conducted within 2 months prior to the referenced 1L therapy start date; (2) their initial follow-up scan was performed at least 28 days after therapy start date; (3) documentation indicated that the follow-up scan was compared to the baseline scan; (4) documentation existed for lesion measurements on both the baseline and follow-up scans; and (5) all measured and non-measured lesions were specifically described in each imaging report. Although RECIST can be applied to multiple imaging modalities, CT is preferred for evaluation of lung cancer [11]. Therefore, we focused on chest CT and PET/CT radiology reports.

Experiment 2: Comparison of Alternative Abstraction Methods for Assessing Real-World Cancer Progression

Three approaches for determining real-world cancer progression were used to abstract outcomes data from individual charts: (1) a radiology-anchored approach; (2) a clinician-anchored approach; and (3) a combination of both. The radiology-anchored approach, conceptually similar to RECIST, focused on identifying new lesions and quantifying tumor burden changes observed by imaging, but was optimized for the type of radiological sources routinely found in EHR data (e.g., “Impression” section within radiology reports). Reports from any imaging modality were considered. The clinician-anchored approach was conceptually focused on the clinician as synthesizer of signals from the entire patient’s chart. The combined approach considers both radiology reports and the clinician’s synthesis: the date of progression is determined by the earliest source document for the event. Definitions of each approach, evidence sources evaluated, and the logic for calculating progression dates are described in Table 1.
Table 1

Non-RECIST-based approaches to determining cancer progression using EHR data (experiment 2)

Radiology-anchored approachaClinician-anchored approachaCombined approacha
DefinitionDocumented in the radiology report as progression based on the radiologist’s interpretation of the imagingDocumented in a clinician’s note as cancer progression based on the clinician’s interpretation of the entire patient chart, including diagnostic procedures and testsDocumented in either the radiology report or clinician note as cancer progression
Primary and corroborating evidence sourcesPrimary: radiology reports

Primary: clinician notes

Corroborating: radiology reports

Primary: radiology reports and clinician notes
Cancer progression dateDate of the first radiology report that indicated a progression eventDate of the first radiology report referenced by the clinician assessment when available, or date of clinician note when no corresponding radiology was conducted or documentedThe earliest report date of the available sources of documentation of a progression event

Three different abstraction approaches for determining real-world cancer progression: (1) radiology-anchored approach, (2) clinician-anchored approach, and (3) combined approach. For each abstraction approach, the approach definitions, source evidence evaluated, and progression date assignment rules are described

aEach approach also considered pathology reports as potential evidence for progression. However, there were no instances of pathology reports preceding radiology reports in the cohort analyzed for this study. In addition, there were no instances of conflicting information between radiology and pathology reports. For simplicity of presentation, pathology reports were excluded from this description as a potential evidence source

Non-RECIST-based approaches to determining cancer progression using EHR data (experiment 2) Primary: clinician notes Corroborating: radiology reports Three different abstraction approaches for determining real-world cancer progression: (1) radiology-anchored approach, (2) clinician-anchored approach, and (3) combined approach. For each abstraction approach, the approach definitions, source evidence evaluated, and progression date assignment rules are described aEach approach also considered pathology reports as potential evidence for progression. However, there were no instances of pathology reports preceding radiology reports in the cohort analyzed for this study. In addition, there were no instances of conflicting information between radiology and pathology reports. For simplicity of presentation, pathology reports were excluded from this description as a potential evidence source

Analysis

For experiment 1, descriptive statistics (frequencies and percentages) were calculated for patients meeting RECIST feasibility criteria. Results were summarized for the percentage of cases that met all criteria, under both the strictest definition (explicit target lesion mention) and the more lenient definition. The saturation sampling completeness threshold, a qualitative research method, was 75% (i.e., at least 75% of patients had sufficient data to allow radiology reports for the RECIST approach to be usable). For experiment 2, demographics, clinical and tumor characteristics, and treatment types were summarized using quantitative descriptive statistics (medians and interquartile ranges or frequencies and percentages, as appropriate). Progression events within the first 14 days of 1L therapy start were excluded from all calculations as they occurred too early to reflect treatment effectiveness. To compare the pre-defined abstraction approaches (Fig. 1), the proportions of patients with a progression event with accompanying 95% confidence intervals (CIs) were calculated. The concordance of progression events and associated dates were also assessed for each approach. Discordant cases were reviewed to determine the discrepancy sources. Among patients with at least one progression event, we assessed frequency (95% CIs) of near-term (within 15 days prior to treatment initiation and within 60 days post progression date), clinically relevant downstream events: death, start of new therapy line (second or subsequent lines), or therapy stop. RWD-based time-to-event analyses for PFS (rwPFS) and time to progression (rwTTP) as well as overall survival (OS) were performed using standard Kaplan–Meier methods [14]. The index date was defined as the date of 1L therapy start, and the event date was defined as the date of the first progression event (rwTTP); the first progression event or date of death (rwPFS); or the date of death due to any cause (OS). In cases with inter-abstractor disagreement on the event date, the date from the abstractor who documented the event first was used in the Kaplan–Meier analysis. Patients lost to follow-up or without the relevant event by the end of the study period were censored at the date of last confirmed activity (last clinical visit or drug administration date). Medians, 95% CIs, and Kaplan–Meier curves were calculated for each outcome. To assess the correlation between rwPFS and OS, the cohort was restricted to patients with a documented death event (N = 123). For the correlation between rwTTP and OS, the cohort was further restricted to patients with documentation of both progression and death events (N = 112 for the clinician anchored approach; 113 for the radiology anchored approach and for the combined approach). Spearman’s rank correlation coefficient (Spearman’s rho) was used for all correlation analyses. The reliability of progression abstraction was assessed by calculating inter-rater agreement for the first progression event in a random subset of patients (N = 55) abstracted in duplicate (two independent abstractors per chart). Patients may have had any number of progression events within or across multiple lines of therapy, but the inter-rater agreement analysis included only up to the first event. Agreement was calculated at patient level and considered from multiple perspectives: (1) agreement on the presence or absence of at least one progression event (event agreement), regardless of date; (2) in cases where both abstractors found at least one progression event (cases with event disagreement or no events were excluded), agreement on when the progression occurred (date agreement); and, (3) a combined approach where both absence or presence of a progression event and the event date, if one was found, contribute toward agreement (overall agreement). Date and overall agreement were calculated for exact date, as well as for 15- and 30-day windows between dates. Open-ended qualitative feedback was collected from abstractors regarding the usability, feasibility, and consistency of each approach.

Results

Experiment 1 evaluated a cohort of 26 aNSCLC patients (Fig. 2). With a strict RECIST definition (radiologist-defined target lesions required in imaging report), no patient chart (0%) yielded data suitable for assessing cancer progression. With more lenient criteria (no explicit mention of target lesions required), only 15 charts (58% of experiment 1 cohort) had radiology reports with descriptions appropriate for RECIST assessment. Of these, only 8 (31% of experiment 1 cohort) directly compared all measured key lesions between two time points, as required by RECIST. Even fewer (6, 23% of experiment 1 cohort) had evidence that all non-measured key lesions were followed between the two time points, also a RECIST component. Completeness did not reach the saturation sampling threshold of 75% completeness: 0–23%, depending on requirement for explicit documentation of key lesions. As a result of infeasibility (lack of usability), output from the RECIST abstraction approach was not further analyzed or compared.
Fig. 2

Assessing applicability of RECIST for defining cancer progression in real-world EHR data in experiment 1. Twenty-six patient charts were randomly selected from the overall cohort of 7584 patients with at least 2 clinical visits and 2 lines of therapy (LoT). RECIST criteria were applied and the numbers of patients meeting the various criteria were recorded

Assessing applicability of RECIST for defining cancer progression in real-world EHR data in experiment 1. Twenty-six patient charts were randomly selected from the overall cohort of 7584 patients with at least 2 clinical visits and 2 lines of therapy (LoT). RECIST criteria were applied and the numbers of patients meeting the various criteria were recorded

Experiment 2: Comparison of Different Abstraction Methods for Assessing Real-World Cancer Progression

In experiment 2 (N  = 200), the median age at advanced diagnosis was 66 years, half were women, 63% were stage IV at diagnosis, most had non-squamous cell histology (73%), and most had a smoking history (85%) (Table 2).
Table 2

Demographic and clinical characteristics of experiment 2 cohort

VariableN = 200
Demographics
 Median age at advanced diagnosis, years [IQR]65.5 [57.0; 72.0]
 Age at advanced diagnosis, n (%)
   < 55 years37 (18.0)
  55–64 years58 (29.0)
  65+105 (52.5)
 Gender, n (%)
  Female100 (50.0)
  Male100 (50.0)
 Race/ethnicity, n (%)
  White137 (68.5)
  Black or African American15 (7.5)
  Asian6 (3.0)
  Other race15 (7.5)
  Unknown/missing27 (13.5)
 Region, n (%)
  Northeast62 (31.0)
  Midwest36 (18.0)
  South67 (33.5)
  West35 (17.5)
Clinical characteristics
 Stage at diagnosis, n (%)
  Stage I13 (6.5)
  Stage II8 (4.0)
  Stage III44 (22.0)
  Stage IV125 (62.5)
  Not reported10 (5.0)
 Histology, n (%)
  Non-squamous cell carcinoma145 (72.5)
  Squamous cell carcinoma46 (23.0)
  NSCLC histology NOS9 (4.5)
 Smoking status, n (%)
  History of smoking169 (84.5)
  No history of smoking25 (12.5)
  Unknown/not documented6 (3.0)
 First-line therapy class, n (%)
  Platinum-based chemotherapy combinations103 (51.5)
  Anti-VEGF-based therapies48 (24.0)
  Single agent chemotherapies28 (14.0)
  EGFR TKIs18 (9.0)
  Non-platinum-based chemotherapy combinations1 (0.5)
  PD-1/PD-L1-based therapies1 (0.5)
  Clinical study drug-based therapies1 (0.5)
 Treatment setting, n (%)
  Community194 (97.0)
  Academic6 (3.0)
 Median follow-up time from advanced diagnosis, months [IQR]13 [9.0; 21.0]
Demographic and clinical characteristics of experiment 2 cohort For most patients (n  = 173, 87% of experiment 2 cohort) both radiology- and clinician-anchored approaches identified at least one progression event. In addition, 7 patients (4% of experiment 2 cohort) had radiology-anchored progression events without confirmation in a clinician note, for a total 180 patients with a radiology-anchored progression event (90% of experiment 2 cohort). Of those 7 cases without clinician confirmation, chart review determined that the discrepancy could be attributed to (1) the radiology report was the last available document in the chart and the clinician assessment was inexistent or unavailable in the chart (5/7), signifying possible death, pending clinical visit, hospice referral, or loss to follow-up, or (2) there was disagreement between the radiology report and the assessment in the clinician note (2/7), (e.g., radiology report recorded progression but the clinician determined stable disease). For the 173 patients with a progression event identified by both radiology- and clinician-anchored approaches, progression dates matched between the two approaches for 156 patients (90%). Among the 17 patients (10%) with differing progression event dates, the radiology-anchored approach identified an earlier progression date in almost all cases (n  = 16, 93%). In all 16 cases, the clinician recorded stable disease immediately following the radiology-anchored progression event but subsequently documented a progression event. In the one final discordant case, the clinician-anchored progression event did not have a corresponding scan documented in the chart. A near-term, downstream event (treatment stop, treatment change, or death, Table 3) was present for 124/173 (71.7%) and 121/180 (67.2%) of patients with cancer progression identified by the clinician-anchored approach and the radiology-anchored approach, respectively. The combined abstraction approach using the first clinician- or radiology-anchored event showed identical results to the radiology-anchored approach.
Table 3

Likelihood of predicting downstream events in experiment 2

Abstraction approach
Radiology-anchoredClinician-anchoredCombined
Number of patients with at least one progression event, n (% of experiment 2 cohort)180 (90.0%)173 (86.5%)180 (90.0%)
Number of patients with a downstream eventa121124121
Proportion of patients with an associated downstream event,  % (95% CI)67.2 (60–74)71.7 (65–79)67.2 (60–74)

aClinically relevant downstream events defined as death, start of new therapy line (second or subsequent lines), or therapy stop. Downstream events occurred 15 days prior and up to 60 days after the progression date

Likelihood of predicting downstream events in experiment 2 aClinically relevant downstream events defined as death, start of new therapy line (second or subsequent lines), or therapy stop. Downstream events occurred 15 days prior and up to 60 days after the progression date Clinician-anchored median rwPFS (Table 4) was higher than radiology-anchored (5.5 [95% CI 4.6–6.3] vs. 4.9 months [95% CI 4.2–5.6]). A similar pattern was observed for rwTTP (Fig. 3). Median OS for all patients was 17 months (95% CI 14–19). Correlations between rwPFS or rwTTP and OS (Table 4) were very similar across abstraction approaches and slightly higher for rwTTP (Spearman’s rho 0.70; 95% CI 0.59–0.78) than for rwPFS (Spearman’s rho [95% CI] ranged from 0.65 [0.53–0.74] to 0.66 [0.55–0.75], depending on approach).
Table 4

Correlation between rwPFS or rwTTP and OS in experiment 2

Abstraction approach
Radiology-anchoredClinician-anchoredCombined
rwPFS
 Median, months (95% CI)4.9 (4.2–5.6)5.5 (4.6–6.3)4.9 (4.2–5.6)
 Correlation with OS, % (95% CI)a,b65 (53–74)66 (55–75)65 (53–74)
rwTTP
 Median, months (95% CI)5.0 (4.2–6.1)5.6 (4.8–6.5)5.0 (4.2–6.1)
 Correlation with OS, % (95% CI)a,c70 (59–78)70 (59–78)70 (59–78)

aSpearman’s rho

bIncludes only patients with an observed death (n = 123)

cIncludes only patients with an observed death and a cancer progression event preceding death (n = 112 for the clinician-anchored approach; n = 113 for the radiology-anchored and the combined approach)

Fig. 3

rwPFS, rwTTP, and OS in experiment 2. Kaplan–Meier estimate curves for overall survival and a progression-free survival (PFS) or b time to progression (TTP), for all three non-RECIST abstraction approaches

Correlation between rwPFS or rwTTP and OS in experiment 2 aSpearman’s rho bIncludes only patients with an observed death (n = 123) cIncludes only patients with an observed death and a cancer progression event preceding death (n = 112 for the clinician-anchored approach; n = 113 for the radiology-anchored and the combined approach) rwPFS, rwTTP, and OS in experiment 2. Kaplan–Meier estimate curves for overall survival and a progression-free survival (PFS) or b time to progression (TTP), for all three non-RECIST abstraction approaches Inter-rater agreement on presence or absence of a first progression event was 96–98%, depending on approach (Table 5). Overall agreement was also similar for all approaches, 71–73%, when 30-day windows between progression event dates were allowed. When considering only cases where both abstractors agreed that at least one progression event occurred (n = 48, 49), the progression event dates were within 30 days of each other in 69–71% of patients.
Table 5

Inter-rater agreement reliability in experiment 2

ApproachAgreement level N Agreement, % (95% CI)
Exact15-day window30-day window
Radiology-anchoredEvent5598 (94–100)
Date4961 (47–75)67 (54–80)69 (56–82)
Overall5564 (51–77)69 (57–81)71 (59–83)
Clinician-anchoredEvent5596 (91–100)
Date4860 (46–74)67 (54–80)71 (58–84)
Overall5562 (49–75)67 (55–79)71 (59–83)
CombinedEvent5598 (94–100)
Date4961 (47–75)69 (56–82)71 (58–84)
Overall5564 (51–77)71 (59–83)73 (61–85)

Patient charts were abstracted in duplicate by different abstractors and agreement (95% CIs) is reported. Event agreement is based on the presence or absence of at least one cancer progression event. Date agreement is based on when the progression occurred, and only calculated in cases where both abstractors recorded a cancer progression. Overall agreement is based on a combined approach where both the absence or presence of a progression event and the date of the event, if one was found, contribute toward agreement

Inter-rater agreement reliability in experiment 2 Patient charts were abstracted in duplicate by different abstractors and agreement (95% CIs) is reported. Event agreement is based on the presence or absence of at least one cancer progression event. Date agreement is based on when the progression occurred, and only calculated in cases where both abstractors recorded a cancer progression. Overall agreement is based on a combined approach where both the absence or presence of a progression event and the date of the event, if one was found, contribute toward agreement Abstractors reported preferring the clinician-anchored approach because it afforded a more comprehensive context. One abstractor reported that the clinician-anchored approach “provide[s] a true comparison of the scan results and guide to overall patient treatment.” Abstractors also reported the clinician-anchored approach to be faster and more straightforward compared to the radiology-anchored approach.

Discussion

EHR-based real-world research requires validated approaches to abstract outcomes in order to glean meaningful insights and ensure consistent descriptions across different data sets and studies. What we really want to know about an intervention is whether patients are doing “better,” “worse,” or “unchanged.” Many proxies have been proposed to answer this question. In prospective clinical trials, RECIST tumor-size measurements on imaging are a common metric. However, it was unknown how clinical trial endpoint language translates to assessments of outcomes documented during routine clinical care. Additionally, EHR outcome abstraction needs to facilitate, rather than hinder, research on large, contemporary real-world patient cohorts. Therefore, we evaluated one RECIST-based and three non-RECIST-based approaches to identify a feasible method suitable for large-scale abstraction of cancer progression from EHRs. We found that it is not feasible to use RECIST (standard or lenient) to abstract cancer progression from EHRs as outcomes are missing at least 75% of the time. This finding is unsurprising given that RECIST application in clinical trial settings often requires enormous resources. While quantitative methods were used exclusively for experiment 2, we determined infeasibility (lack of usability) for experiment 1 using the saturation sampling method, a common qualitative approach [16, 17] that enables early stopping if additional cases are unlikely to yield a positive outcome. Feasibility requires adequate completeness of data; however, at 20 patients the RECIST approach failed to reach a very lenient 75% completeness threshold. Further enrichment to 26 patients showed persistently poor data completeness. We then assessed three non-RECIST approaches to define cancer progression from the EHR using technology-enabled human curation (Table 1); all had more complete endpoint information than the RECIST-based approach, yielded similar analytic results, and identified progression events at a frequency consistent with expected clinical benchmarks in this disease based on clinical studies. Progression events were identified in 86–90% of cases, regardless of approach, and predicted near-term downstream clinical events, such as treatment change and death, more than two-thirds of the time. For progression events without a near-term, downstream clinical event, chart review confirmed that these were due to clinical care realities (e.g., few subsequent treatment options or a treatment “holiday”). Regardless of abstraction approach, rwP performed as expected in time-to-event analyses; a median rwPFS of approximately 5 months is similar to that observed in published aNSCLC clinical trials [18, 19], and correlations between rwPFS/rwTTP and OS were consistent with, or higher than, results from clinical trials [18, 20, 21]. Given our reliance on RECIST in clinical trials, one might anticipate greater reliability from the radiology-anchored abstraction approach; however, the clinician-anchored approach was similarly reliable and abstractor feedback indicated that it may be more scalable. The similar reliability may be due to the overlap between clinician and radiology assessments of disease progression in real-world settings. Modest rwPFS differences for the clinician- and radiology-anchored approaches likely reflect differences in the underlying conceptual structure. The radiology-anchored approach identified more progression events, but the clinician-anchored approach was more likely to be associated with a treatment change or death. This finding suggests that the clinician-anchored approach may benefit from clinician notes that synthesize information and adjudicate potentially conflicting findings. In the radiology-anchored approach abstractors are directed to consider radiology reports as the primary source of truth; however, radiologists may lack access to EHRs to support interpretation of imaging findings. Median rwPFS was approximately 2 weeks shorter for the radiology-anchored approach when compared with the clinician-anchored approach, likely because imaging often precedes a clinician visit and/or assessment in real-world settings. Nonetheless, modest differences and overlapping CIs preclude strong conclusions. Importantly, the association between rwPFS and OS in this cohort showed no meaningful differences between approaches. Despite these similar quantitative results, abstractor feedback differed. The clinician-anchored approach was favored, as a clinician synthesis resonated with the broader context of the patient’s clinical journey. In addition, abstractors reported shorter abstraction times for the clinician-anchored approach. These features support the feasibility and scalability of abstracting progression from the EHR with a clinician-anchored approach. In contrast, the radiology-anchored approach is potentially less scalable, at least for community oncology medical records (most patients in this cohort). Further, it is plausible that the radiology-anchored approach may not perform as well when determination of progression is based on non-radiologic findings (e.g., cutaneous progression, symptom history, biomarker evidence), and in resource- or access-limited settings. Overall, the clinician-anchored cancer progression abstraction method is the most practical of those studied for abstraction of tumor-based endpoints in EHR data. A scalable approach can support research on a large patient population with a rapid turnaround time, generating contemporary data to answer questions of relevance to current and future patients. There are several limitations to consider in this study. There are several sources of potential subjectivity (e.g., radiologists, clinicians, and abstractors). However, rwPFS correlates with OS in ways similar to RECIST-based PFS in clinical trial studies, and the correlation between the radiology- and clinician-anchored approaches was similar. We implemented clear instructions and abstractor training to mitigate subjectivity during the abstraction process. Although similar inter-rater reliability was observed across approaches, further training and experience could improve abstractor reliability. Another limitation of this analysis is the grouping of patients who presented with aNSCLC (only treated with palliative intent) with those who initially presented with an earlier stage of NSCLC and progressed to aNSCLC after curative intent treatment. Correlations between OS and intermediate endpoints may vary by these subgroups; future studies could stratify by stage at initial presentation. In addition, future research with patients diagnosed in 2018 or later (when the AJCC 8th edition staging system came into effect, and after our study enrollment ended) could evaluate the potential impact of this staging update [22, 23]. More broadly, there may be limitations when applying this work to other diseases and settings. Although real-world studies in other high-prevalence tumor types, such as breast cancer, have analyzed intermediate endpoints also based on clinician assessments [24-26], cross-tumor validation studies are lacking. Tumor, treatment, and assessment differences (e.g., tumor kinetics, tumor response pattern, assessment cadence and modality, role of biomarker data, and the availability and type of subsequent treatment options) vary across tumor types and may affect the performance of these endpoint variables. Development of a set of rules to reconcile conflicting progression data and integrate additional data sources may help address these differences. Our work predominantly included EHR data generated in the community oncology setting. Since management and documentation patterns may vary across treatment settings, additional exploration focused on academic centers may help ascertain generalizability. Whether this clinician-anchored approach will hold true for other real-world endpoints such as tumor response requires further examination. Lastly, any approach utilizing time-dependent endpoints is susceptible to bias due to data missingness or the frequency of assessments, particularly if unbalanced across treatment groups. Further study is required to characterize the extent of potential bias.

Conclusions

This study lays the foundations for developing and validating surrogate endpoints of cancer progression suitable for generating high-quality RWD based on information routinely found in the EHR. Identification of a practical clinician-anchored abstraction method for characterizing the response of real-world cancer patients to treatment enhances the relevance and interpretability of RWE in oncology. On the basis of the study findings, we foresee two important next steps. In order to fully harness the potential value of EHR data, outcome measures tailored to the unique features of the data source are needed. Once those outcome measures are developed, such as the methodology for abstraction of cancer progression described here, a comprehensive validation framework must be established to enable robust and standardized characterization of such approaches. Second, we need to tailor and expand this outcomes development and validation approach to different contexts and cancer types. It is also important to explore the broader opportunities to assess other outcome metrics suitable for the real-world setting such as quality of life. Below is the link to the electronic supplementary material. Supplementary material 1 (DOCX 18 kb)
  20 in total

1.  Treatment patterns and clinical outcomes among patients receiving palbociclib in combination with an aromatase inhibitor or fulvestrant for HR+/HER2-negative advanced/metastatic breast cancer in real-world settings in the US: Results from the IRIS study.

Authors:  Gavin Taylor-Stokes; Debanjali Mitra; John Waller; Katie Gibson; Gary Milligan; Shrividya Iyer
Journal:  Breast       Date:  2018-10-20       Impact factor: 4.380

2.  Comparative effectiveness research: a report from the Institute of Medicine.

Authors:  Harold C Sox; Sheldon Greenfield
Journal:  Ann Intern Med       Date:  2009-06-30       Impact factor: 25.391

3.  Opportunities and challenges in leveraging electronic health record data in oncology.

Authors:  Marc L Berger; Melissa D Curtis; Gregory Smith; James Harnett; Amy P Abernethy
Journal:  Future Oncol       Date:  2016-03-08       Impact factor: 3.404

4.  End points and United States Food and Drug Administration approval of oncology drugs.

Authors:  John R Johnson; Grant Williams; Richard Pazdur
Journal:  J Clin Oncol       Date:  2003-04-01       Impact factor: 44.544

Review 5.  The Eighth Edition Lung Cancer Stage Classification.

Authors:  Frank C Detterbeck; Daniel J Boffa; Anthony W Kim; Lynn T Tanoue
Journal:  Chest       Date:  2016-10-22       Impact factor: 9.410

6.  New response evaluation criteria in solid tumours: revised RECIST guideline (version 1.1).

Authors:  E A Eisenhauer; P Therasse; J Bogaerts; L H Schwartz; D Sargent; R Ford; J Dancey; S Arbuck; S Gwyther; M Mooney; L Rubinstein; L Shankar; L Dodd; R Kaplan; D Lacombe; J Verweij
Journal:  Eur J Cancer       Date:  2009-01       Impact factor: 9.162

Review 7.  Endpoints for assessing drug activity in clinical trials.

Authors:  Richard Pazdur
Journal:  Oncologist       Date:  2008

8.  Treatment Patterns and Outcomes of Patients With Metastatic ER+/HER-2- Breast Cancer: A Multicountry Retrospective Medical Record Review.

Authors:  Samantha K Kurosky; Debanjali Mitra; Giovanni Zanotti; James A Kaye
Journal:  Clin Breast Cancer       Date:  2017-10-14       Impact factor: 3.225

9.  Prediction of survival benefits from progression-free survival benefits in advanced non-small-cell lung cancer: evidence from a meta-analysis of 2334 patients from 5 randomised trials.

Authors:  Silvy Laporte; Pierre Squifflet; Noémie Baroux; Frank Fossella; Vassilis Georgoulias; Jean-Louis Pujol; Jean-Yves Douillard; Shinzohy Kudoh; Jean-Pierre Pignon; Emmanuel Quinaux; Marc Buyse
Journal:  BMJ Open       Date:  2013-03-13       Impact factor: 2.692

10.  Development and Validation of a High-Quality Composite Real-World Mortality Endpoint.

Authors:  Melissa D Curtis; Sandra D Griffith; Melisa Tucker; Michael D Taylor; William B Capra; Gillis Carrigan; Ben Holzman; Aracelis Z Torres; Paul You; Brandon Arnieri; Amy P Abernethy
Journal:  Health Serv Res       Date:  2018-05-14       Impact factor: 3.402

View more
  36 in total

1.  Effectiveness of First-line Immune Checkpoint Blockade Versus Carboplatin-based Chemotherapy for Metastatic Urothelial Cancer.

Authors:  Emily Feld; Joanna Harton; Neal J Meropol; Blythe J S Adamson; Aaron Cohen; Ravi B Parikh; Matthew D Galsky; Vivek Narayan; John Christodouleas; David J Vaughn; Rebecca A Hubbard; Ronac Mamtani
Journal:  Eur Urol       Date:  2019-07-28       Impact factor: 20.096

2.  Status Update on Data Required to Build a Learning Health System.

Authors:  Monica M Bertagnolli; Brian Anderson; Kelly Norsworthy; Steven Piantadosi; Andre Quina; Richard L Schilsky; Robert S Miller; Sean Khozin
Journal:  J Clin Oncol       Date:  2020-03-25       Impact factor: 44.544

Review 3.  Contemporary use of real-world data for clinical trial conduct in the United States: a scoping review.

Authors:  James R Rogers; Junghwan Lee; Ziheng Zhou; Ying Kuen Cheung; George Hripcsak; Chunhua Weng
Journal:  J Am Med Inform Assoc       Date:  2021-01-15       Impact factor: 4.497

4.  Antiproliferative Effects of Telotristat Ethyl in Patients with Neuroendocrine Tumors: The TELEACE Real-World Chart Review Study.

Authors:  Michael A Morse; Eric Liu; Vijay N Joish; Lynn Huynh; Mu Cheng; Mei Sheng Duh; Kiernan Seth; Pablo Lapuerta; David C Metz
Journal:  Cancer Manag Res       Date:  2020-07-30       Impact factor: 3.989

5.  Uptake and Survival Outcomes Following Immune Checkpoint Inhibitor Therapy Among Trial-Ineligible Patients With Advanced Solid Cancers.

Authors:  Ravi B Parikh; Eun Jeong Min; E Paul Wileyto; Fauzia Riaz; Cary P Gross; Roger B Cohen; Rebecca A Hubbard; Qi Long; Ronac Mamtani
Journal:  JAMA Oncol       Date:  2021-12-01       Impact factor: 31.777

6.  Influence of first-line chemotherapy regimen on survival outcomes of patients with advanced urothelial carcinoma who received second-line immune checkpoint inhibitors.

Authors:  Benjamin Miron; Elizabeth Handorf; Kevin Zarrabi; Matthew R Zibelman; Fern Anari; Pooja Ghatalia; Elizabeth R Plimack; Daniel M Geynisman
Journal:  Urol Oncol       Date:  2022-07-02       Impact factor: 2.954

7.  Automated NLP Extraction of Clinical Rationale for Treatment Discontinuation in Breast Cancer.

Authors:  Matthew S Alkaitis; Monica N Agrawal; Gregory J Riely; Pedram Razavi; David Sontag
Journal:  JCO Clin Cancer Inform       Date:  2021-05

8.  Effectiveness of Alpelisib + Fulvestrant Compared with Real-World Standard Treatment Among Patients with HR+, HER2-, PIK3CA-Mutated Breast Cancer.

Authors:  Stuart Turner; Stephen Chia; Hemanth Kanakamedala; Wei-Chun Hsu; Jinhee Park; David Chandiwana; Antonia Ridolfi; Chu-Ling Yu; Juan Pablo Zarate; Hope S Rugo
Journal:  Oncologist       Date:  2021-05-13

9.  Older adult participation in cancer clinical trials: A systematic review of barriers and interventions.

Authors:  Mina S Sedrak; Rachel A Freedman; Harvey J Cohen; Hyman B Muss; Aminah Jatoi; Heidi D Klepin; Tanya M Wildes; Jennifer G Le-Rademacher; Gretchen G Kimmick; William P Tew; Kevin George; Simran Padam; Jennifer Liu; Andrew R Wong; Andrea Lynch; Benjamin Djulbegovic; Supriya G Mohile; William Dale
Journal:  CA Cancer J Clin       Date:  2020-10-01       Impact factor: 508.702

10.  Real-world effectiveness of post-trastuzumab emtansine treatment in patients with HER2-positive, unresectable and/or metastatic breast cancer: a retrospective observational study (KBCSG-TR 1917).

Authors:  Takahiro Nakayama; Tetsuhiro Yoshinami; Hiroyuki Yasojima; Nobuyoshi Kittaka; Masato Takahashi; Shoichiro Ohtani; Seung Jin Kim; Hiroyuki Kurakami; Naoko Yamamoto; Tomomi Yamada; Takehiko Takata; Norikazu Masuda
Journal:  BMC Cancer       Date:  2021-07-09       Impact factor: 4.430

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.