Literature DB >> 30037859

Accuracy of administrative databases in detecting primary breast cancer diagnoses: a systematic review.

Iosief Abraha1,2, Alessandro Montedori1, Diego Serraino3, Massimiliano Orso1,2, Gianni Giovannini1, Valeria Scotti4, Annalisa Granata5, Francesco Cozzolino1, Mario Fusco5, Ettore Bidoli3.   

Abstract

OBJECTIVE: To define the accuracy of administrative datasets to identify primary diagnoses of breast cancer based on the International Classification of Diseases (ICD) 9th or 10th revision codes.
DESIGN: Systematic review. DATA SOURCES: MEDLINE, EMBASE, Web of Science and the Cochrane Library (April 2017). ELIGIBILITY CRITERIA: The inclusion criteria were: (a) the presence of a reference standard; (b) the presence of at least one accuracy test measure (eg, sensitivity) and (c) the use of an administrative database. DATA EXTRACTION: Eligible studies were selected and data extracted independently by two reviewers; quality was assessed using the Standards for Reporting of Diagnostic accuracy criteria. DATA ANALYSIS: Extracted data were synthesised using a narrative approach.
RESULTS: From 2929 records screened 21 studies were included (data collection period between 1977 and 2011). Eighteen studies evaluated ICD-9 codes (11 of which assessed both invasive breast cancer (code 174.x) and carcinoma in situ (ICD-9 233.0)); three studies evaluated invasive breast cancer-related ICD-10 codes. All studies except one considered incident cases.The initial algorithm results were: sensitivity ≥80% in 11 of 17 studies (range 57%-99%); positive predictive value was ≥83% in 14 of 19 studies (range 15%-98%) and specificity ≥98% in 8 studies. The combination of the breast cancer diagnosis with surgical procedures, chemoradiation or radiation therapy, outpatient data or physician claim may enhance the accuracy of the algorithms in some but not all circumstances. Accuracy for breast cancer based on outpatient or physician's data only or breast cancer diagnosis in secondary position diagnosis resulted low.
CONCLUSION: Based on the retrieved evidence, administrative databases can be employed to identify primary breast cancer. The best algorithm suggested is ICD-9 or ICD-10 codes located in primary position. TRIAL REGISTRATION NUMBER: CRD42015026881. © Author(s) (or their employer(s)) 2018. Re-use permitted under CC BY-NC. No commercial re-use. See rights and permissions. Published by BMJ.

Entities:  

Keywords:  accuracy; administrative database; breast cancer; sensitivity and specificity; systematic review; validity

Mesh:

Year:  2018        PMID: 30037859      PMCID: PMC6059263          DOI: 10.1136/bmjopen-2017-019264

Source DB:  PubMed          Journal:  BMJ Open        ISSN: 2044-6055            Impact factor:   2.692


Based on a prepublished protocol, this is the first review that systematically addressed the accuracy of administrative databases in identifying subjects with breast cancer. We performed a comprehensive electronic databases search complemented with reference check of relevant articles, and we evaluated the quality of reporting of included studies by the Standards for Reporting of Diagnostic checklist. We considered only papers written in English and this might have introduced a language bias. The knowledge and experience of the International Classification of Diseases (ICD)-9/ICD-10 coders could have influenced the quality of breast cancer case definition in each study, and consequently the results presented in our review could be biased by this factor. Generalisability of validated administrative databases is limited to the context in which they are generated.

Introduction

The burden of cancer is increasingly growing among populations, and it is associated with major economic expenditure worldwide,1 especially in low-income and middle-income countries.2 As breast cancer is the most common cancer and the leading cause of cancer death in women,3 knowledge of its epidemiology and the ability to monitor related outcomes over time is important for health planning services. Administrative healthcare databases are increasingly being used in oncology for epidemiological evaluation,4 population outcome research,5 drug utilisation reviews,6–8 evaluation of health service delivery and quality9 10 as well as health policy development.11–13 Generally, these databases gather longitudinal information concerning health resource utilisation regarding hospitalisations, outpatient care and, often, drug prescriptions and vital statistics.14 In other words, these databases provide a readily available source of ‘real-world’ data on a large population of unselected patients allowing the performance of less expensive and more representative assessment of disease surveillance and outcome research compared with randomised trials.15 16 By definition, administrative healthcare databases contain data that are routinely and passively collected without an a priori research question, as they are usually established for billing or, in general, for administrative purposes, and not for research uses. Hence, the diagnostic codes used to identify, for example, cancers, must be validated according to an accepted ‘reference standard’ diagnosis.17 In validation studies of administrative databases, the reference standard usually used is the clinical chart or cancer registry.18 The current International Classification of Diseases, 9th revision, (ICD-9) codes are 233.0 for breast carcinoma in situ and 174.0–174.9 for invasive breast cancer, whereas the ICD-10 codes are D05.0-D05.9 and C50.0-C50.9, respectively. These codes help to identify subjects that have breast cancer within an administrative healthcare database. Since the clinical diagnosis of breast cancer is based on a combination of clinical and/or instrumental examinations and a pathological assessment,19 these codes are limited in confirming whether a specific subject within the databases truly has the disease of interest. As a result, researchers have proposed a number of different claim-based algorithms for case identification of breast cancers, such as a combination of healthcare claims data,20 the use of chemotherapy21 and the number of medical claims on separate dates.11 In addition, since patients with metastatic cancer have different prognoses and typically different treatment patterns to those with earlier-stage malignancies, researchers suggest using algorithms to identify patients with metastatic cancer.11 22 To our knowledge, data on the validity of breast cancer diagnosis codes have not been synthesised in the medical literature. Our objective was to determine the best algorithms with which to identify breast cancer cases using administrative databases based on a comprehensive systematic search of primary studies that validated ICD-9 or ICD-10 codes related to breast cancer. The present work has been conceived within a project of validating three large administrative healthcare databases in Italy concerning ICD-9-CM codes for breast, colorectal and lung cancers.23–25 For our purposes, it was important to identify all available case definitions or algorithms that best identify subjects with the cancer diseases of interest as outlined in our protocol.26

Methods

This study is part of different projects supported by national and local funding with the objectives of assessing case definitions of diseases as well as validating ICD-9 codes for cancer23 26 and other diseases.27–29 As outlined in the protocol,26 the target population consisted of patients with primary diagnosis of breast cancer, the index test was represented by administrative data algorithms related to breast cancer, the reference standard was represented by medical charts, validated electronic health records or cancer registries.

Literature search

Comprehensive searches of MEDLINE, EMBASE, Web of Science and the Cochrane Library from their inception to April 2017 were performed to identify published peer-reviewed literature. We developed a search strategy based on the combination of: (a) keywords and Medical Subject Heading (MeSH) terms to identify records concerning breast cancer; (b) terms to identify studies likely to contain validity or accuracy measures and (c) a search strategy designed to capture studies that used healthcare administrative databases based on the combination of terms used by Benchimol et al30 and the Mini-Sentinel’s program.31 32 The developed search strategy is reported in the online supplementary file 1. To retrieve additional articles, the authors searched relevant reference lists of key articles. Titles and abstracts were screened for eligibility by two independent reviewers. Discrepancies were solved by discussion. This systematic review was prepared according to the Preferred Reporting Items for Systematic reviews and Meta-Analysis Protocols (PRISMA-P) 2015 Statement33 and the results were presented following the PRISMA flow diagram (figure 1).34 A protocol of this review has also been published at the BMJ Open26 as well as an outline in the PROSPERO International Prospective Register of systematic reviews with registration number CRD42015026881 (http://www.crd.york.ac.uk/PROSPERO).
Figure 1

Study screening process.

Study screening process.

Inclusion criteria

Full texts of eligible peer-reviewed articles without publication date restriction, published in English that used administrative data for the ICD-9 or ICD-10 codes related to breast cancer diagnoses were obtained. For each study, the following inclusion criteria were applied: (a) the presence of a reference standard (clinical chart, cancer registry or electronic health records), together with the presence of any case definition or algorithm for breast cancer; (b) the presence of at least one test measure (eg, sensitivity, positive predictive value (PPV), etc); (c) the data source was from an administrative database (ie, a database in which data are routinely and passively collected without an a priori research question) and (d) the study database was from a representative sample of the general population. We aimed to focus on primary diagnosis of breast cancers, hence studies that considered algorithms to identify cancer history, cancer progression or recurrence were not evaluated. In addition, studies that considered index test databases that were not truly administrative (eg, cancer registries, epidemiology surveillance systems, etc) were excluded. However, studies that used electronic health records to validate breast cancer were also included.35 36

Selection process

After screening titles and abstracts, we subsequently obtained full texts of eligible articles to determine if they meet the inclusion and exclusion criteria. We conducted data abstraction using standardised data collection forms that were tested on a sample of three eligible articles. Two review authors working independently and in duplicate were involved in titles and abstracts screening, full-texts screening and data abstraction (FC, MO, AG, VS). Discrepancies were resolved by consensus, and where necessary with the involvement of a third review author (IA). Calibration exercises were performed at each level of the process.

Data extraction

Data extraction included the following information: the details of the included study (including title, year and journal of publication, country of origin and sources of funding; the type of disease (invasive, in situ or both); the target population from which the administrative data were collected; the type of administrative database used (eg, hospitalisation discharge data), outpatient records (eg, physician billing claims); the ICD-9 or ICD-10 codes used or the administrative data algorithms tested (including Current Procedural Terminology; prescription fills, etc); the position of the ICD codes in the discharge abstract database (ICD codes in primary position indicate the principal diagnosis, that is, the condition identified at the end of the admission, which is the main cause of the need for treatment or diagnostic investigations; ICD codes in secondary positions refer to secondary diagnoses, that is, conditions that coexist at the time of the admission or which develop after that time and which influence the treatment received and/or the length of hospital stay); the modality of development of the algorithm (eg, using Classification and Regression Trees, logistic regression, expert opinion, etc); external validation; use of training and testing cohorts; the reference standard used to determine the validity of the diagnostic codes (eg, medical chart review, patient self-reports, cancer registry, etc); the characteristic of the test used to determine the validity of the diagnostic code or algorithm (eg, sensitivity, specificity, PPVs and negative predictive values (NPVs), area under the receiver operating characteristic curve, likelihood ratios and kappa statistics).

Quality assessment

The design and methods of the included primary studies were assessed using a checklist developed by Benchimol et al,30 based on the criteria published by the Standards for Reporting of Diagnostic accuracy (STARD) initiative for the accurate reporting of studies using diagnostic studies.37 The checklist is provided in online supplementary file 2. The presence of potential biases within the studies were reported in a descriptive way.

Analysis

For each algorithm, we abstracted the performance statistics provided in the included studies including sensitivity, specificity, PPV and NPV. Where necessary, we calculated validation statistics together with their 95% CIs as far as raw numbers for cases and controls were provided.

Patient and public involvement

Patients and the public were not directly involved. This was a retrospective study based on the consultation of electronic medical literature.

Results

After removing duplicate records identified through MEDLINE, EMBASE, Web of Science and The Cochrane Library, 2929 citations were screened in titles and abstracts. Overall, we assessed 41 full-text articles for eligibility, of which 1710 38–53 were included in the final evaluation. In addition, a reference check of pertinent articles permitted the identification of five potentially relevant studies of which four were included in the final analysis54–57 (figure 1). The list of excluded studies, together with the reasons of their exclusion is reported in the online supplementary file 3.

Study characteristics

The included studies were published between 1992 and 2015 and collected data between 1977 and 2011. Fifteen studies were performed in the USA,39 41 42 44–47 51–53 55–59 two were conducted in Italy,10 38 two in France,40 54 one in Japan49 and one in Australia.43 Seventeen studies used Cancer Registry data as the reference standard,38–40 42 43 45–47 49–57 60 four studies used medical chart review.45 50 51 57 Eighteen studies10 38 39 41 42 44–48 50–57 evaluated ICD-9 codes, and three studies40 43 49 evaluated ICD-10 codes. Of the studies that evaluated ICD-9 codes, 11 reported the evaluation of the ICD-9 codes related to invasive breast cancer (code 174.x) and carcinoma in situ (ICD-9 233.0)10 38 42 44–47 50–53; three evaluated only invasive cancer (ICD-9 174.x)39 41 55; three studies did not specify the number of the ICD-9 codes evaluated.54 56 57 Three studies evaluated ICD-10 codes for invasive breast cancer without evaluating carcinoma in situ codes.40 43 49 In terms of representativeness or generalisability, the studies varied greatly. Eight studies considered all women beneficiaries of the Medicare programme, USA), age 65 years or above residing in specific areas,39 42 46 51–53 55 56 nine studies considered all women with any age38 49 or aged 15+54 or 20+10 38 40 45 50 57 in specific areas, two studies considered all women aged 40+44 or 45+43 residing in specific areas and three studies randomly sampled residents at a national level,41 or residents at regional level.47 59 Basic characteristics of these studies are displayed in table 1.
Table 1

Characteristics of included studies

First author, year of publicationPeriod of data collectionCountryRecords evaluated (N)Source populationType of administrative dataDiagnostic codesAlgorithmReference standard
Fisher et al41 19921984–1985USA33 cases (any position); 24 cases (first position) from each 239 hospitalsAll National Medical beneficiaries,Medicare claim database: inpatient hospital discharge.174–174.9(a) Diagnosis in any position. (b) Diagnosis in primary position.Medical records review
McBean et al55 19941986–1987USA5744 casesPersons 65 years of age and older living in the five states participating in the SEER Program.Medicare claim database: inpatient hospital data.174–174.9New cases of breast cancer with diagnostic codes in any position.Cancer Registry (SEER)
Solin et al50 19941988–1989USA469 casesWomen aged ≥21 years enrolled in the USA. Healthcare in the southeastern Pennsylvania region.Claims database included inpatient hospital stays, short procedure unit stays and professional services. Each claim included the USA (comprised ICD-9 code and CPT-4).174–174.9; 233.0(A) Initial algorithm based on new case of breast carcinoma+one or more of the following: (1) mastectomy, (2) partial mastectomy with lymphadenectomy, (3) excision, breast biopsy or partial mastectomy+lymphadenectomy, (4) excision, breast biopsy or partial mastectomy+the diagnosis of carcinoma of the breast, (5) excision, breast biopsy or partial mastectomy followed by radiation therapy treatment or (6) excision, breast biopsy or partial mastectomy followed by chemotherapy treatment. (B) Best algorithm with multiple modifications of the initial algorithm.Medical records review
Warren et al53 19961989USA3454 casesAll women aged 65+ years with one or more hospitalisations with a diagnosis of breast cancer in Medicare.Medicare Hospital Inpatient (ICD-9-CM).174–174.9; 233.0(i) Any hospitalisation breast cancer (ICD-9-CM 174–174.9 and 233.0) as principal diagnosis. (ii) Incident cases of breast cancer (no prior hospitalisation with breast cancer in any of the 5 positions from 1984 to 1988 or history of breast cancer (ICD-9-CM V.10.3) appearing as any of the five positions from 1984 to 1989). (iii) Analysis limited to women who were residents in one of the five SEER states.Cancer Registry (SEER)
Solin et al51 19971993–1994USA177 casesAll women aged ≥65 years enrolled in the US Healthcare (Pennsylvania and New Jersey).Claims database included hospital inpatient, short procedure unit stays and professional services (ICD-9 diagnosis code, and the CPT-4 procedure code).174–174.9; 233.0This study was performed to evaluate prospectively a previously published algorithm50 used to identify women with the new diagnosis of carcinoma of the breast.Medical records review
McClish et al46 19971986–1989USA3690 casesAll residents aged 65+ years diagnosed with breast cancer (Virginia).MEDPAR (Medicare) inpatient hospital claim database (ICD- 9 CM).174–174.9; 233.0Incident cases of breast cancer ICD-9-CM 174; V174.9; 233 and 233.0.Cancer Registry
Cooper et al61 19991984–1993USA71 862 casesAll women aged 64+ years with breast cancer (Atlanta, Detroit, Seattle-Puget Sound, San Francisco Oakland, Connecticut, Hawaii, Iowa, New Mexico and Utah).(1) MEDPAR (Medicare: inpatient hospital claim database (ICD-9 CM and specific procedural codes ICD-9-CM and HCPCS/CPT-4. (2) Part B: physician and outpatient claim data.174–174.9Basic: incidence breast cancer (174.0–174.9); (i) all other inpatient diagnostic codes; inpatient cancer-specific surgical code (local excision/lumpectomy: 85.20, 85.21, 85.22); Part B: first position diagnosis code; any other part B diagnosis codes and part B cancer-specific surgical code; (ii) first position diagnostic coding+the inclusion of the following: all other part B diagnosis codes; part B cancer-specific procedural code; inpatient first position diagnosis; all other inpatient diagnoses and inpatient cancer-specific surgery.Cancer Registry (SEER)
Warren et al52 19991992USAWomen residing in the SEER states n=6 59 260; cases=6784.All Medicare eligible women residing in one of five SEER states who were age 65 years and older as of 1 January 1992.Medicare inpatient and physician claim database (ICD-9).174–174.9; 233.0Model 1: inpatient primary diagnosis (174–174.9, 233.0), excluding prevalent cases), inpatient-secondary diagnosis, and physician bills—reference group). Model 2: inpatient-secondary diagnosis and cases identified from the physician data, breast cancer-related procedures.Cancer Registry (SEER)
Leung et al45 19991994–1996USA1033 casesAll women aged 21 years or older who were enrolled in Health Net (California).Claims database includes claims received for inpatient hospital stays, short procedure unit stays and professional services (code ICD-9-CM).174–174.9; 233.0Basic breast cancer diagnosis and one of the following: (1) mastectomy; (2) partial mastectomy with lymphadenectomy; (3) excision, breast biopsy or partial mastectomy+lymphadenectomy; (4) excision, breast biopsy or partial mastectomy+diagnosis of carcinoma; (5) excision, breast biopsy or partial mastectomy followed by radiation therapy or (6) excision, breast biopsy or partial mastectomy followed by chemotherapy.Medical chart review
Freeman et al42 20001990–1992USA7464 cases; 1415 controls:Females aged 65–74 years (in 1992) diagnosed with breast cancer (San Francisco/Oakland, Detroit, Atlanta and Seattle and the states of Connecticut, Iowa, New Mexico, Utah and Hawaii).ICD-9 inpatient record; outpatient record; physician claim (Medicare).174–174.9; 233.0; V103Model 1 (6 predictors): hospital inpatient: breast cancer principal or additional diagnosis) or 1992 hospital. Model 2: (10 predictors) (model 1 or breast cancer-related procedure (mastectomy, partial mastectomy, excisional biopsy, incisional biopsy). Model 3 (36 predictors): (hospital inpatient, hospital outpatient, physician claim): model 1 and breast cancer-related procedure (mastectomy biopsy, biopsy, chemotherapy); mammography, breast cancer-related radiology, radiation oncology, laboratory test on a hospital outpatient. Model 4 (hospital inpatient, hospital outpatient, physician claim): model 1 and breast cancer-related procedure (mastectomy biopsy, biopsy, chemotherapy); mammography, breast cancer-related radiology, radiation oncology, laboratory test on a hospital outpatient; biopsy, radiation oncology, laboratory test on a physician claim; mammography, other radiological procedures in 1992 on a hospital outpatient claim.Cancer Registry (SEER)
Wang et al57 20011989–1991USA8872 casesAll women aged 20 years and older who were enrolled in either Medicaid or Medicare and PAAD (New Jersey State).Medicaid in patient files.ICD-9 code: not reportedNew cases of breast cancer: primary algorithm definitions:: surgical claims for (a) a CPT or ICD-9 procedure code for a mastectomy; (b) a CPT or ICD-9 procedure code for an excision of a breast mass+CPT or ICD-9 procedure code for an axillary node biopsy; (c) a DRG hospitalisation code for breast cancer surgical hospitalisation. Alternative algorithms: combinations of ICD-9 diagnostic codes, DRG hospitalisation codes, and codes for non-surgical treatments.Cancer registry
Koroukian et al44 20031997–1998USA2635 incident casesWomen aged 40 years or older (Ohio).Medicaid claims and enrolment files. ICD-9-CM.174–174.9; 233.0Incident of breast cancer (ICD-9 174.0–174.9 233.0) and combinations of diagnosis and procedures codes (chemotherapy or radiation therapy, mastectomy, lumpectomy).Cancer Registry (OCISS)
Ganry et al54 20031998France198 incident casesAll women aged 15 years or older who were diagnosed or treated (in the Amiens University Hospital and five general hospitals) of the Somme area.French hospital database adapted from the Diagnosis Related Group (DRG).ICD-9 code: not reportedNew case of breast cancer—at least one of the following criteria: (a) breast cancer as primary diagnosis, alone or with (i) mastectomy; (ii) partial mastectomy with lymphadenectomy or (iii) excision, breast biopsy or partial mastectomy for procedures; (b) breast cancer as secondary diagnosis, with (i) chemotherapy as principal diagnosis or (ii) without specific procedures (excluding prevalent cases: women with history of breast cancer between 1991 and 1997).Cancer registry (French Somme Area)
Nattinger et al47 2004Validation set: 1994; training set: 1995USA7607 cases and 120 317 controlsTraining set: claims from 7700 SEER-Medicare breast cancer subjects (age 65+ years) diagnosed in 1995, and 124 884 controls. Validation set: claims from 7607 SEER-Medicare breast cancer subjects diagnosed in 1994, and 120 317 controls.Random sample; ICD-9 inpatient record; outpatient record; physician claim (Medicare).174–174.9; 233.0Four-step algorithm: (1) breast cancer diagnosis+procedure code; (2) both conditions: (a) mastectomy or a lumpectomy or partial mastectomy followed by at least one outpatient or carrier claim for radiotherapy with a breast cancer diagnosis; (b) at least two outpatient or carrier claims on different dates containing breast cancer as the primary diagnosis; (3) this step applies to all potential cases that passed step 1, but were not directly included at step 2; (4) removal of prevalent cases.Cancer Registry (SEER)
Penberthy et al59 20051995USA249 casesWomen aged 65+ years with breast cancer diagnosis.(a) lnpatient Medicare; (b) inpatient or Part B claims.ICD-9 174Six case definitions: A1) diagnosis first position; A2) inpatient diagnosis in any position; A3) inpatient diagnosis in any position+inpatient surgical procedure; B1) inpatient diagnosis in any position+inpatient surgical procedure; B2) inpatient diagnosis in any position+inpatient surgical procedure OR a diagnostic procedure+diagnosis+a surgery or chemotherapy or radiation therapy procedure in an outpatient or physician office record within 4 months of a diagnostic procedure; B3) inpatient diagnosis in any position OR a diagnosis+a surgery or chemotherapy or radiation therapy procedure in an outpatient or physician office record.Cancer Registry (Virginia State); medical chart review
Setoguchi et al56 20071997–2000USA2004 casesSubjects aged 65+ years Medicare recipients in Pennsylvania.Medicare inpatient hospital claim and drug benefit programme data.UnclearFour algorithms based on the combination of the following: (a) ICD-9 diagnosis codes (numbers not provided), (b) CPT codes for screening procedures, surgical procedures, radiation therapy, chemotherapy and nuclear medicine procedures and/or (c) National Drug Code prescription codes for medications used for cancer treatment available in Pharmaceutical Assistance Contract for the Elderly.Cancer Registry (Pennsylvania State)
Baldi et al38 20082000 training set, 2001 validation setItaly925 casesAll residents in Piedmont region.Regional inpatient administrative database.174–174.9; 233.0Algorithms based on combination between (i) ICD-9-CM diagnosis breast cancer (invasive 174.0–174.9, in situ 233.0); and ICD-9-CM procedures code incisional breast biopsy: 85.12 excision or destruction of breast tissue: 85.20–85.25 subcutaneous mastectomy: 85.33–85.36 mastectomy: 85.41–85.48Cancer Registry (Piedmont Region)
Couris et al40 20092002France995 casesWomen aged 20 years or older living in one of the nine French districts covered by a cancer registry in 2002.Inpatient hospital administrative data (French National Institute of Statistics and Economic Studies) based on ICD-10.C50.0 to C50.9(a) Principal diagnosis for invasive breast cancer—ICD-10 codes C50.0 to C50.9; (b) principal diagnosis+specific surgery procedures.French cancer registries
Yuen et al10 20112002–2005Italy11 615 casesWomen aged 20 years older with incident breast cancer (Emilia-Romagna region).Regional administrative database (Hospital discharge files).174–174.9; 233.0Women having a diagnosis code for cancer as well as a principal or secondary surgical code for lumpectomy or mastectomy: principal or secondary procedure indicating excision or destruction of breast tissue (ICD-9-CM code 85.20–85.25) or mastectomy (ICD-9-CM code 85.41–85.48); and principal or secondary diagnosis of carcinoma in situ of the breast (ICD-9-CM code 233.0) or malignant neoplasm of the breast (ICD-9-CM code 174.0–174.9).Cancer registry (AIRTUM)
Kemp et al43 20132004–2008Australia2039 women with invasive breast tumourWomen aged 45+ years who had completed breast cancer-related items in the baseline survey of the 45 and up study (New South Wales).i) Administrative hospital separations records (ICD-10-AM); ii) outpatient medical service claims; iii) prescription medicines claims and iv) the 45 and up study baseline survey.C50.0 to C50.9Principal inpatient diagnosis of invasive breast cancer using ICD-10- AM codes C50.0-C50.9.Cancer Registry
Sato et al49 20152011Japan50 056 women included in the study cohort (633 with breast cancer)Women with no prior cancer-related history, from the claims data at a single institution between 1 January and 31 December 2011.ICD for oncology, third edition (ICD-O-3): topography code of breast cancer (C500 to C506, C508, C509).C50.0 to C50.914 definitions starting from (1) breast cancer alone and subsequent addition of (2) diagnosis code related to breast cancer (3) diagnostic imaging code (4) biopsy code (5) marker test code (6) surgery code (7) chemotherapy code (8) medication code (9) radiation procedure code (10) the other code related to breast cancer (11) diagnosis code related to breast cancer or marker test code (12) surgery, chemotherapy, medication or radiation procedure code (13) diagnosis code related to the breast cancer, marker test code, surgery, chemotherapy, medication or radiation procedure code (14) ≥3 diagnoses of breast cancer.Cancer Registry

AIRTUM, Associazione Italiana dei Registri Tumori; CPT-4: current procedural terminology-4; HCPCS, Healthcare Common Procedure Coding System; ICD, International Classification of Diseases; MEDPAR, Medicare Annual Demographic Files, the Medicare Provider Analysis and Review; OCISS, Ohio Cancer Incidence Surveillance System; SEER, Surveillance, Epidemiology, and End Results Program.

Characteristics of included studies AIRTUM, Associazione Italiana dei Registri Tumori; CPT-4: current procedural terminology-4; HCPCS, Healthcare Common Procedure Coding System; ICD, International Classification of Diseases; MEDPAR, Medicare Annual Demographic Files, the Medicare Provider Analysis and Review; OCISS, Ohio Cancer Incidence Surveillance System; SEER, Surveillance, Epidemiology, and End Results Program.

Validity of breast cancer data

Accuracy results by initial algorithms

All the studies considered new (incident) breast cancer cases except Fisher et al.41 Nineteen studies presented the initial accuracy results based on breast cancer diagnosis only; in 18 studies the diagnosis was in primary position,10 38–50 52–54 56 in 1 study in any position55 and in 2 studies the position was unclear,51 57 whereas 2 studies evaluated breast cancer diagnosis with surgical procedures.10 38 Sensitivity was reported by 17 studies, and was at least 80% in 65% (n=11) of them10 41 43 44 46 47 49 53–55 57 (range 57%–99%). PPV, obtained from 19 studies, was ≥83% in the majority (n=14) of them (range 15%–98%). Specificities resulted higher than 98% in all the eight studies that provided sufficient data to permit calculation.10 40 43 47 49 52 54 56 Similarly, the NPV for the five studies for which it was possible to calculate was ≥99%.40 43 52 54 56 Table 2 displays the results of the algorithm with which the studies presented their initial data stratified by ICD codes.
Table 2

Accuracy results by initial algorithm in the 21 included studies

Study IDInitial algorithmSensitivity (95% CI)Specificity (95% CI)PPV (95% CI)NPV (95% CI)
ICD-9
 Fisher et al41 1992BCD in primary position96 (79 to 100)88 (70 to 98)
 McBean et al55 1994BCD in any position9796
 Solin et al50 1994BCD in primary position88 (85 to 91)
 Warren et al53 1996BCD in primary position94 (93 to 95)9783 (78 to 89)
 McClish et al46 1997BCD in primary position83 (82 to 84)91 (90 to 93)
 Solin et al51 1997BCD (unclear position)83 (77 to 87)
 Leung et al45 1999BCD in primary position84 (82 to 87)
 Warren et al52 1999BCD in primary position57 (55 to 59)99 (99 to 99)91 (90 to 93)99 (98 to 100)
 Cooper et al61 1999BCD in primary position68 (68 to 69)
 Freeman et al42 2000BCD in primary position68 (66 to 70)74 (72 to 76)
 Wang et al57 2001BCD (unclear position)89 (88 to 90)
 Koroukian et al44 2003BCD in primary position69 (66 to 71)15 (13 to 17)
 Ganry et al54 2003BCD in primary position85 (80 to 90)100 (100 to 100)98 (94 to 99)100 (100 to 100)
 Nattinger et al47 2004BCD in primary position80 (79 to 81)100 (100 to 100)89 (87 to 92)
 Penberthy et al59 2005BCD in primary position5396
 Setoguchi et al56 2007≥1 BCD in primary position87 (86 to 89)100 (100 to 100)50 (49 to 52)100 (100 to 100)
 Baldi et al38 2008BCD in primary position+surgical procedures74 (71 to 77)90 (87 to 92)
 Yuen et al10 2011BCD in primary position+surgical procedures85 (84 to 86)99 (99 to 99)91 (90 to 91)
ICD-10
 Couris et al40 2009BCD in primary position69 (66 to 72)99 (99 to 99)57 (54 to 60)100 (100 to 100)
 Kemp et al43 2013BCD in primary position86 (85 to 88)99 (99 to 99)86 (84 to 87)100 (100 to 100)
 Sato et al49 2015BCD in primary position99 (98 to 100)99 (93 to 99)66 (63 to 69)

BCD, breast cancer diagnosis; ICD, International Classification of Disease; NPV, negative predictive value; PPV, positive predictive value.

Accuracy results by initial algorithm in the 21 included studies BCD, breast cancer diagnosis; ICD, International Classification of Disease; NPV, negative predictive value; PPV, positive predictive value.

Accuracy results by combinations of diagnosis and surgical procedures

Twelve studies reported validation results using algorithms with different combinations.10 38 40 43–45 49–52 54 56 All algorithms, except in two studies,10 38 started evaluating basic breast cancer codes and progressively added surgical procedures, secondary diagnosis, chemotherapy and/or radiotherapy. The addition of one or more of these items to the algorithms produced different results over the basic accuracy results obtained with the use of the diagnosis code alone. The addition of excision to the incident diagnosis of invasive cancer codes did not add any value to the PPV in the studies by Solin et al50 (88% vs 89%), Leung et al45 (83% vs 84%) and Kemp et al43; conversely, in the study by Solin et al51 while in the first algorithm there were no improvements between the new diagnosis and the addition of the excision (83% vs 84%), using the best algorithm set, the PPV rose from 84% to 92% when excision was added to the basic breast diagnosis code. In the study by Koroukian et al,44 the addition of mastectomy or mastectomy/lumpectomy significantly raised the PPV from 15% to 84% and 87%, respectively. In the study by Kemp et al,43 the sensitivity and PPV values of breast cancer diagnosis remained substantially unchanged with the addition of mastectomy, lumpectomy or both (PPV 86% vs 89%). Setoguchi et al56 proposed four algorithms: the algorithm based on one or more diagnoses of breast cancer generated a sensitivity of 87% and a PPV of 50%; the addition of any surgical procedure lowered the sensitivity to 46% but enhanced the PPV to 82%. In the study by Sato et al,49 the addition of any code related to breast cancer, marker tests, surgical procedures, chemotherapy treatment or radiation therapy did not affect the sensitivity (that resulted high: 98%) but raised the PPV from 66% to 83%. Similarly in the study by Ganry et al,54 the addition of any breast or lymph nodal surgical procedures enhanced the PPV from 91% to 98%. Table 3 shows the sensitivities and PPVs with the respective CIs of the studies that in addition to accuracy measures of breast cancer diagnosis also reported accuracy data of surgical procedures.
Table 3

Results of studies validating diagnoses of breast cancer (first row) and surgical procedures (subsequent rows)

#Author/yearAlgorithms invasive breast cancerSensitivitySpecificityPPVNPV
1Solin et al50 1994Diagnosis incident cases88 (85 to 91)
2Solin 1994Mastectomy95 (92 to 98)
3Solin 1994Partial mastectomy with lymphadenectomy96 (91 to 100)
4Solin 1994Excision and lymphadenectomy100 (100 to 100)
1Solin et al51 1997Initial algorithm: diagnosis incident cases83 (78 to 89)
2Solin 1997Initial algorithm: mastectomy95 (90 to 100)
3Solin 1997Initial algorithm: partial mastectomy with lymphadenectomy95 (88 to 100)
4Solin 1997Initial algorithm: excision and lymphadenectomy92 (83 to 100)
5Solin 1997Best algorithm: diagnosis incident cases84 (79 to 90)
6Solin 1997Best algorithm: mastectomy82 (76 to 88)
7Solin 1997Best algorithm: partial mastectomy with lymphadenectomy84 (78 to 89)
8Solin 1997Best algorithm: excision and lymphadenectomy84 (78 to 89)
1McClish et al46 1997incident cases identified in MEDPAR83 (82 to 84)
2McClish 1997incident cases identified in VCR82 (81 to 83)
3McClish 1997aggregated (VCR+MEDPAR)97 (96 to 97)
4McClish 1997MEDPAR definitive surgical therapy80 (79 to 81)
5McClish 1997VCR definitive surgical therapy87 (86 to 88)
1Leung et al45 1999Initial algorithm: diagnosis84 (82 to 87)
2Leung 1999Mastectomy92 (90 to 95)
3Leung 1999Partial mastectomy with lymphadenectomy98 (96 to 100)
4Leung 1999Excision, breast biopsy or partial mastectomy plus lymphadenectomy92 (87 to 98)
1Cooper et al61 1999First set of analysis (increase in SE including in order inpatient other diagnosis, surgical, part B, etc): inpatient, first position diagnostic codes68 (68 to 69)
2Cooper 1999First set of analysis: inpatient, surgical79 (79 to 79)
3Cooper 1999First set of analysis: part B, first position91 (91 to 91)
4Cooper 1999First set of analysis: part B, surgical94 (93 to 94)
5Cooper 1999Second set of analysis (increase in SE including in order part B other diagnosis, surgical, inpatient, etc): part B, first position66 (66 to 66)
6Cooper 1999Second set of analysis: part B, other diagnostic codes77 (77 to 77)
7Cooper 1999Second set of analysis: part B, surgical81 (81 to 81)
8Cooper 1999Second set of analysis: inpatient, first position91 (91 to 91)
9Cooper 1999Second set of analysis: inpatient, surgical94 (93 to 94)
1Freeman et al42 2000Primary diagnosis: hospital inpatient in Medicare Provider Analysis (MEDPAR)68 (66 to 70)74 (72 to 76)
2Freeman 2000Mastectomy hospital inpatient53 (51 to 56)73
3Freeman 2000Partial mastectomy hospital inpatient7 (4 to 11)64
4Freeman 2000Excisional biopsy hospital inpatient8 (5 to 12)56
5Incisional biopsy hospital inpatient8 (5 to 11)73
1Ganry et al54 2003Hospitalisation with breast cancer as primary diagnosis: mastectomy100 (100 to 100)
2Ganry 2003Hospitalisation with breast cancer as primary diagnosis: partial mastectomy with lymphadenectomy100 (100 to 100)
3Ganry 2003Hospitalisation with breast cancer as primary diagnosis: biopsy/excision plus the diagnosis of carcinoma100 (100 to 100)
1Kemp et al43 2013Diagnosis of invasive breast cancer86 (85 to 88)100 (100 to 100)86 (84 to 87)100 (100 to 100)
2Kemp 2013Lumpectomy61 (59 to 63)99 (99 to 99)52 (50 to 54)99 (99 to 99)
3Kemp 2013Mastectomy33 (31 to 35)100 (100 to 100)71 (68 to 74)99 (99 to 99)
4Kemp 2013Lumpectomy OR mastectomy84 (83 to 86)99 (99 to 99)56 (55 to 58)100 (100 to 100)

MEDPAR, Medicare Annual Demographic Files, the Medicare Provider Analysis and Review; NPV, negative predictive value; PPV, positive predictive value; VCR, Virginia Cancer Registry.

Results of studies validating diagnoses of breast cancer (first row) and surgical procedures (subsequent rows) MEDPAR, Medicare Annual Demographic Files, the Medicare Provider Analysis and Review; NPV, negative predictive value; PPV, positive predictive value; VCR, Virginia Cancer Registry.

Accuracy results by combinations of diagnosis and surgical procedures followed by chemoradiation or radiation therapy

Six studies added chemotherapy or radiation therapy procedures to their algorithm.44 45 49–51 54 Compared with the initial algorithm with only the diagnosis of breast cancer, the PPV value increased in all instances to values higher than 94% in four studies.45 50 51 54 However, in all studies except one the algorithms contained surgical procedures. Table 4 displays the sensitivity and PPV values for the studies that combined chemoradiation or radiation therapy procedures with diagnosis of breast cancer.
Table 4

Results of studies that combined surgical procedures followed by chemoradiation or radiation therapy with diagnosis of breast cancer

NAuthor/yearAlgorithms invasive breast cancerSensitivity % (CI)Specificity % (CI)PPV % (CI)
1Solin et al50 1994Diagnosis incident cases88 (85 to 91)
2Solin 1994Excision followed by radiation therapy94 (88 to 99)
3Solin 1994Excision followed by chemotherapy94 (88 to 100)
1Solin et al51 1997Initial algorithm: diagnosis incident cases83 (78 to 89)
2Solin 1997Initial algorithm: excision followed by radiation therapy treatment97 (91 to 100)
3Solin 1997Initial algorithm: excision followed by chemotherapy90 (77 to 100)
4Solin 1997Best algorithm: diagnosis incident cases84 (79 to 90)
5Solin 1997Best algorithm: excision followed by radiation therapy treatment84 (78 to 89)
6Solin 1997Best algorithm: excision followed by chemotherapy84 (78 to 89)
1Leung et al45 1999Initial algorithm: diagnosis 84 (82 to 87)
2Leung 1999Excision, breast biopsy or partial mastectomy followed by radiation therapy96 (94 to 98)
3Leung 1999Excision, breast biopsy or partial mastectomy followed by chemotherapy93 (90 to 97)
1Koroukian et al44 2003Incident breast cancer15 (13 to 17)
2Koroukian 2003Breast cancer diagnosis, chemotherapy or radiation therapy34 (29 to 39)
3Koroukian 2003Breast cancer diagnosis, lumpectomy, chemotherapy or radiation therapy85 (78 to 92)
1Ganry et al54 2003Hospitalisation with breast cancer (primary diagnosis): without any procedure91 (81 to 100)
2Ganry 2003Hospitalisation with breast cancer (secondary diagnosis): chemotherapy as primary diagnosis98 (94 to 100)
1Sato et al49 2015Diagnosis of breast cancer99 (99 to 100)99 (93 to 100)66 (63 to 69)
2Sato 2015Diagnosis of breast cancer+diagnosis code related to the breast cancer, marker test code, surgery, chemotherapy, medication or radiation procedure code97 (96 to 100)100 (100 to 100)83 (80 to 85)

PPV, positive predictive value.

Results of studies that combined surgical procedures followed by chemoradiation or radiation therapy with diagnosis of breast cancer PPV, positive predictive value.

Accuracy results based on the position of the diagnosis

Three studies provided results based on the position of the diagnosis.41 52 54 Fisher et al41 provided sensitivity and PPV for breast cancer diagnosis in any position and it resulted in similar results in the primary position (sensitivity 97% and PPV 84% in any position; sensitivity 96% and PPV 88% in the primary position). Accuracy results for secondary position breast cancer diagnosis was provided by two studies and the estimates resulted lower than the accuracy results for diagnosis in primary position in the studies by Ganry et al54 and Warren et al.52 PPVs were 26% and 65% in secondary position against 91% and 91% in primary position, respectively (see online supplementary file 4, eTable 1).

Accuracy results based on outpatient or physician’s data

Only two studies assessed the accuracy of breast cancer diagnosis codes based on outpatient or physician’s records42 52; the other 19 studies considered inpatient data alone or in combination with other types of data (short procedure unit stays, professional services, prescription medicines claims, etc). For the physician’s mammography and laboratory data, the sensitivities resulted 87% in both cases but with very low corresponding PPV values (0% and 15%, respectively).42 The remaining cases concerning biopsy, surgical procedures, nodal dissection in the physician records or outpatient records showed very low PPVs (see online supplementary file 4, eTable 2).

Stratified analysis by administrative data source, type of ICD code, country of origin and publication year

Accuracy data stratified by setting of diagnosis showed that outpatient accuracy data were much lower than diagnosis in primary position, although the outpatient accuracy data were reported by only one study.42 In terms of codes, both ICD-9 and ICD-10 showed significant variation in both sensitivity and PPV. In terms of country of origin, most of the studies were conducted in the USA where the variability of the accuracy results showed important variation. The studies conducted in Italy10 38 and France40 54 showed similar ranges of sensitivities, whereas the studies conducted in Italy performed better in terms of PPVs than any other country. Accuracy results of the initial algorithm did not change over time and the range of sensitivities remained similar between the studies published before 2001 compared with the studies published after 2000. PPVs estimates remained also similar provided one outlier, that is, the study by Koroukian et al,44 is excluded. Table 5 shows ranges of sensitivities and PPVs stratified by administrative data source, type of ICD code, country of origin and publication year.
Table 5

Range of sensitivities and PPVs stratified by administrative data source, type of ICD code and country of origin

Range of sensitivitiesRange of PPVs
Administrative data source
 Inpatient (primary position only)53%–99% (18 studies)15%–98% (19 studies)
 Outpatient (outpatient diagnosis only)9% (1 study)19% (1 study)
Type of ICD
 ICD-9 (initial algorithm)53%–97% (15 studies)15%–98% (16 studies)
 ICD-10 (initial algorithm)69%–99% (3 studies)57%–86% (3 studies)
Country of origin
 USA (initial algorithm)53%–97% (12 studies)15%–96% (13 studies)
 Italy (initial algorithm)74%–85% (2 studies)90%–91% (2 studies)
 France (initial algorithm)69%–85% (2 studies)57%–98% (2 studies)
 Japan (initial algorithm)99% (1 study)66% (1 study)
 Australia (initial algorithm)86% (1 study)86% (1 study)
Accuracy over time
 Before 2001 (initial algorithm)57%–97% (7 studies)74%–96% (9 studies)
 After 2000 (initial algorithm)53%–99% (11 studies)15%–98% (10 studies)

ICD, International Classification of Disease; PPV, positive predictive value.

Range of sensitivities and PPVs stratified by administrative data source, type of ICD code and country of origin ICD, International Classification of Disease; PPV, positive predictive value.

Quality of the studies

All the studies explicitly reported their intention to evaluate the accuracy of the administrative database and described validation cohort, age, disease and location of participants. All the studies reported inclusion criteria and only seven45 49–51 55 57 59 (33%) did not report exclusion criteria, and three52 53 56 (14%) did not report any description regarding the patient sampling method. In terms of the methodology used, all the studies described the methods used to calculate diagnostic accuracy, none of the studies described number, training and expertise of persons reading reference standards; none of the studies reported the consistency and the number of persons involved in reading reference standards and, of the studies that used the medical chart review as the reference standard only one41 reported the blinding of the interpreters. In terms of statistical methods, all the studies except one55 described adequately the statistics used to obtain accuracy. None of the studies reported at least four estimates of diagnostic accuracy. The most common statistics used to estimate diagnostic accuracy were sensitivity in 17 studies10 38–41 43 44 46 47 49 52–57 59 (81%), PPV in 19 studies10 38 40–47 49–56 59 (90%) and specificity in 9 studies10 40 43 47 49 52–54 56 (43%); 10 studies10 39 44 46 47 49 50 52 53 57 (48%) reported accuracy results for subgroups; and only 6 studies10 38 41 46 47 49 (29%) reported CIs (see online supplementary file 2).

Discussion

Summary of findings

To our knowledge, this is the first review that systematically addressed the validity of algorithms related to breast cancer diseases in administrative databases. Using several medical literature databases, we have identified a significant number of validation studies related to breast cancer disease. Because of the heterogeneity of the results due to the different settings of each included study, we decided to present them in a descriptive manner rather than aggregate them by means of a meta-analysis. Findings from this review suggest that algorithms based on ICD-9 or ICD-10 codes related to breast cancer are accurate in identifying subjects with invasive breast cancer when the diagnosis is in primary position and the algorithm is based on incident cases. Sixty-seven per cent of the studies reported sensitivities or PPVs higher than 80% for inpatient breast cancer code at the initial presentation. The addition of other fields such as surgical procedures, chemotherapy or radiation therapy, outpatient data and physician claims may improve the accuracy results but depend on the accuracy measure used. Breast cancer codes in secondary position yielded lower accuracy values.

Quality of primary studies and heterogeneity

The overall quality of the studies included in our review was judged quite good. There are only some concerns about the items of the modified STARD checklist related to the description of data collection (who identified the patients, who collected data and whether the authors used an a priori data collection form) of which, evaluating the primary studies, we were not able to find these descriptions in the text. However, we do not think that this could significantly affect the results of our review, as it was common to all the studies and it could be related in general to the peculiarity of the studies validating administrative databases that are substantially different from the typical diagnostic accuracy studies. Regarding reference standard, most of the studies used cancer registries and the confirmation of the cancer disease was based on the presence of the corresponding code within the registry. When medical charts were used as a reference standard, in three studies the diagnosis of carcinoma of the breast was confirmed when there was evidence of a histological documentation of ‘invasive carcinoma of the breast’, ‘intraductal carcinoma (ductal carcinoma in situ) of the breast’ or ‘Paget’s disease of the breast’,45 50 51 whereas Fisher et al41 reported that ‘accredited records technicians, blinded to the coding in the original records, reviewed the medical records, selected the supportable diagnoses and procedures and translated them into ICD-9-CM’. The included studies differed in the geographical area, temporal period, healthcare system, reference standard considered and other factors, and this heterogeneity could explain the variability of the diagnostic accuracy measures. Because of the heterogeneity of the results due to the different settings of each included study, we decided to present them in a descriptive manner rather than aggregate them by means of a meta-analysis.

Prioritising accuracy measures

In assessing the validity of healthcare databases, researchers will need to weigh the relative importance of epidemiological measures and prioritise the accuracy measure that is most important to a particular study. As pointed out by Chuback et al,35 for example, contrary to PPV estimates, sensitivity and specificity do not depend on the disease prevalence but can vary across populations. Privileging sensitivity to specificity is relevant in a scenario where identifying all cases with the characteristic of interest is important rather than only those with severe disease characteristics. PPV is generally preferred when one wants to ensure that only the subjects who truly have the condition of interest are included in the study. In our assessment, 19 studies provided PPV measures, 15 of which also measured sensitivity.10 38 40–44 46–49 53–55 Most of these performed an accuracy assessment with the intent to define the incidence of breast or other cancer diseases. Several authors included different variables in their algorithm, including surgical, chemoradiation or radiation treatment, outpatient care, such as physician’s claims in order to obtain algorithms with a balanced value between PPV and sensitivity. While three studies obtained similar values between sensitivity and PPV,40 43 55 in six studies10 38 42 46 48 54 the initial PPVs were higher than sensitivity and most of these studies had the priority of estimating the incidence of breast cancer disease. Sato et al49 attributed the gain of optimal sensitivity (90%) and PPV (99%) to the use of both inpatient and outpatient claims data. The authors argue that they might have obtained high sensitivity but low PPV if they had used the outpatient database alone. In two Italian validation studies10 38 of regional administrative databases, the combination of hospital diagnosis together with surgical procedures accurately identified the majority of cases in the cancer registry (PPV 90% and 91%, respectively). In other circumstances, the aims of the studies were substantially methodological. Freeman et al,42 who aimed at obtaining the optimal combination of predictors, used a logistic regression model using 1992 data from the linked SEER registries. The authors were able to obtain a high sensitivity (90%) with the use of three kinds of claims data (inpatients, outpatients and physician services), but with a loss of PPV (70%) that was probably due to limitations in distinguishing recurrent and secondary cancers. Cooper et al39 only investigated the sensitivity of diagnostic and procedural coding for case ascertainment of breast and five other cancer diseases. The authors used two sets of analyses: the first set was the inpatient Medicare claims, which include diagnoses and procedures ICD-9 codes; the second set was the part B claims, which include physician and outpatient data. The first set considered the sensitivity of inpatient first position (68.3%) and the increase in sensitivity provided by including additional fields (other diagnosis (76.1%), surgical (79.1%), part B first position (91%), part B other diagnosis (93.1%), part B surgical code (93.6%)). In the second set of analyses, they considered the sensitivity of part B first position (66%) and included the following additional fields: part B other diagnosis (76.9%), part B surgical (80.9%), inpatient first position (90.9%), inpatient other diagnosis (93%) and inpatient surgical (93.6%). Conversely, the aim of Nattinger et al47 was to maintain a high specificity and they proposed a four-step algorithm to identify women with surgically treated incident breast cancer that was applied in both a validation set and a training set. For their objective, they considered cases treated in the ambulatory surgical setting as well as prevalent cases. The authors were able to obtain high specificity (99.9%) with a decrease in sensitivity from 85% to 80% but with a good PPV performance (range 89%–93%). As recognised by the authors, the algorithm may have little usefulness in determining the incidence of breast cancer, but it may be much more relevant for outcome classification.35

Strengths and limitations

Our strengths include the use of comprehensive electronic databases with reference checks of relevant reviews of articles, the use of STARD criteria to assess the quality of reporting of primary studies, transparency based on prepublication of a protocol (online supplementary file 5), the use of detailed and explicit eligibility criteria and the use of duplicate and independent processes for study selection, data abstraction and data interpretation. We must acknowledge that our assessment was focused on primary breast cancer and we did not take into account diagnosis of metastases due to breast cancer. A recent study that evaluated the accuracy of ICD-9 codes in identifying metastatic breast and other cancer diseases found that the performance of the metastases codes from Medicare claims data compared with the gold standard of SEER stage was poor and never exceeded 80% for any of the accuracy measures for any stage for any cancer metastatic disease.60 Other studies reported similar low values of accuracy and this may misclassify a significant number of patients and lead to a biased assessment of survival.22 61 62 Second breast cancer recurrences and second primary breast cancers are of interest in the epidemiological and outcome research of breast cancer. In our assessment, we did not consider studies that used algorithms to identify recurrences and second breast cancer events. A recent study assessed several algorithms to identify second breast cancer events following early stage invasive breast cancer and found high accuracy measures.63 In addition, we were not able to consider articles that were not written in English and this may have introduced language bias. In addition, despite the comprehensive nature of our search, a few pertinent articles may have been missed given that some identified articles did not use the term ‘administrative database’ as a subject heading, and the term is not recognised as a MeSH by Medline. Indeed, we were able to identify four primary studies54–57 using the Cited-By’ tools in PubMed, Google Scholar or checking the reference of included studies. Fourth, the knowledge and experience of the ICD-9/ICD-10 coders could have influenced the quality of breast cancer case definition in each study, and consequently the results presented in our review could be biased by this factor. Finally, we emphasise that the applicability of validation studies depends much on the methods used to identify subjects with the condition of interest to validate the algorithm because this may influence the disease prevalence, and the generalisability of the subjects characteristics as well as the diagnostic accuracy measures. Hence, the generalisability of a database is limited to the setting in which the validation has been performed.64 For example, while Medicare covers the elderly41 42 47 55 59 61 and Medicaid covers indigent and other particular group of patients groups,44 57 the US Healthcare, is an independent practice association model, that may represent patient populations of a relatively higher socioeconomic class.50 51 Hence, inference from these validated databases cannot be made to those who despite residing in the same area of the residents registered in the above reported systems but do not benefit from them or to subjects aged 64 years or less as is in most cases of the Medicare system. Conversely, the database in Italy10 38 and France40 54 where the provision of healthcare is universally provided to residents, the applicability of the results from the validated databases is adequate, although it cannot be extended at a national level. Finally, we found a study that validated data from a single institution in Japan and as acknowledged by the authors it is unclear whether the accuracy results can be directly applicable to other hospitals.49

Conclusion

In summary we conclude that, based on the retrieved evidence, administrative databases can be employed to identify primary breast cancer. The best algorithm suggested is ICD-9 or ICD-10 codes located in primary position. Caution should be used when surgical procedures, chemotherapy, radiation therapy or outpatient data and physician claims are added to the algorithm. We believe that our findings will help researchers that would like to validate breast cancer ICD-9 or ICD-10 codes in administrative databases using either cancer registry or medical charts.
  61 in total

1.  Ability of Medicaid claims data to identify incident cases of breast cancer in the Ohio Medicaid population.

Authors:  Siran M Koroukian; Gregory S Cooper; Alfred A Rimm
Journal:  Health Serv Res       Date:  2003-06       Impact factor: 3.402

2.  Evaluation of an algorithm to identify incident breast cancer cases using DRGs data.

Authors:  O Ganry; A Taleb; J Peng; N Raverdy; A Dubreuil
Journal:  Eur J Cancer Prev       Date:  2003-08       Impact factor: 2.497

3.  Ability of Medicare claims data and cancer registries to identify cancer cases and treatment.

Authors:  D K McClish; L Penberthy; M Whittemore; C Newschaffer; D Woolard; C E Desch; S Retchin
Journal:  Am J Epidemiol       Date:  1997-02-01       Impact factor: 4.897

4.  Treatment patterns and metastasectomy among mCRC patients receiving chemotherapy and biologics.

Authors:  Xue Song; Zhongyun Zhao; Beth Barber; Christopher Gregory; Peter Feng Wang; Stacey R Long
Journal:  Curr Med Res Opin       Date:  2010-12-06       Impact factor: 2.580

5.  The utility of Medicare claims data for measuring cancer stage.

Authors:  G S Cooper; Z Yuan; K C Stange; S B Amini; L K Dennis; A A Rimm
Journal:  Med Care       Date:  1999-07       Impact factor: 2.983

6.  Use of Medicare hospital and physician data to assess breast cancer incidence.

Authors:  J L Warren; E Feuer; A L Potosky; G F Riley; C F Lynch
Journal:  Med Care       Date:  1999-05       Impact factor: 2.983

7.  The sensitivity of Medicare claims data for case ascertainment of six common cancers.

Authors:  G S Cooper; Z Yuan; K C Stange; L K Dennis; S B Amini; A A Rimm
Journal:  Med Care       Date:  1999-05       Impact factor: 2.983

8.  The economic burden of metastatic breast cancer: a U.S. managed care perspective.

Authors:  Alberto J Montero; Sara Eapen; Brian Gorin; Paulette Adler
Journal:  Breast Cancer Res Treat       Date:  2012-06-09       Impact factor: 4.872

9.  Use of ICD-9 coding as a proxy for stage of disease in lung cancer.

Authors:  Simu K Thomas; Sandra E Brooks; C Daniel Mullins; Claudia R Baquet; Sanjay Merchant
Journal:  Pharmacoepidemiol Drug Saf       Date:  2002-12       Impact factor: 2.890

10.  Validity of breast, lung and colorectal cancer diagnoses in administrative databases: a systematic review protocol.

Authors:  Iosief Abraha; Gianni Giovannini; Diego Serraino; Mario Fusco; Alessandro Montedori
Journal:  BMJ Open       Date:  2016-03-18       Impact factor: 2.692

View more
  6 in total

1.  Colorectal Cancer Risk Is Impacted by Sex and Type of Surgery After Bariatric Surgery.

Authors:  Hisham Hussan; Samuel Akinyeye; Maria Mihaylova; Eric McLaughlin; ChienWei Chiang; Steven K Clinton; David Lieberman
Journal:  Obes Surg       Date:  2022-06-22       Impact factor: 3.479

2.  Long-term antimüllerian hormone patterns differ by cancer treatment exposures in young breast cancer survivors.

Authors:  Beth Zhou; Brian Kwan; Milli J Desai; Vinit Nalawade; Kathryn J Ruddy; Paul C Nathan; Henry J Henk; James D Murphy; Brian W Whitcomb; H Irene Su
Journal:  Fertil Steril       Date:  2022-02-23       Impact factor: 7.490

Review 3.  Missing something? A scoping review of venous thromboembolic events and their associations with bariatric surgery. Refining the evidence base.

Authors:  Walid El Ansari; Kareem El-Ansari
Journal:  Ann Med Surg (Lond)       Date:  2020-08-17

4.  Detection of incident breast and colorectal cancer cases from an administrative healthcare database in Catalonia, Spain.

Authors:  J M Escribà; M Banqué; F Macià; J Gálvez; L Esteban; L Pareja; R Clèries; X Sanz; X Castells; J M Borrás; J Ribes
Journal:  Clin Transl Oncol       Date:  2019-10-04       Impact factor: 3.405

5.  Evaluation of the Incidence of Hematologic Malignant Neoplasms Among Breast Cancer Survivors in France.

Authors:  Marie Joelle Jabagi; Norbert Vey; Anthony Goncalves; Thien Le Tri; Mahmoud Zureik; Rosemary Dray-Spira
Journal:  JAMA Netw Open       Date:  2019-01-04

6.  Identifying incident cancer cases in dispensing claims: A validation study using Australia's Repatriation Pharmaceutical Benefits Scheme (PBS) data.

Authors:  B Daniels; H E Tervonen; S-A Pearson
Journal:  Int J Popul Data Sci       Date:  2019-03-19
  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.