Literature DB >> 22822041

Identifying primary and recurrent cancers using a SAS-based natural language processing algorithm.

Justin A Strauss1, Chun R Chao, Marilyn L Kwan, Syed A Ahmed, Joanne E Schottinger, Virginia P Quinn.   

Abstract

OBJECTIVE: Significant limitations exist in the timely and complete identification of primary and recurrent cancers for clinical and epidemiologic research. A SAS-based coding, extraction, and nomenclature tool (SCENT) was developed to address this problem.
MATERIALS AND METHODS: SCENT employs hierarchical classification rules to identify and extract information from electronic pathology reports. Reports are analyzed and coded using a dictionary of clinical concepts and associated SNOMED codes. To assess the accuracy of SCENT, validation was conducted using manual review of pathology reports from a random sample of 400 breast and 400 prostate cancer patients diagnosed at Kaiser Permanente Southern California. Trained abstractors classified the malignancy status of each report.
RESULTS: Classifications of SCENT were highly concordant with those of abstractors, achieving κ of 0.96 and 0.95 in the breast and prostate cancer groups, respectively. SCENT identified 51 of 54 new primary and 60 of 61 recurrent cancer cases across both groups, with only three false positives in 792 true benign cases. Measures of sensitivity, specificity, positive predictive value, and negative predictive value exceeded 94% in both cancer groups. DISCUSSION: Favorable validation results suggest that SCENT can be used to identify, extract, and code information from pathology report text. Consequently, SCENT has wide applicability in research and clinical care. Further assessment will be needed to validate performance with other clinical text sources, particularly those with greater linguistic variability.
CONCLUSION: SCENT is proof of concept for SAS-based natural language processing applications that can be easily shared between institutions and used to support clinical and epidemiologic research.

Entities:  

Mesh:

Year:  2012        PMID: 22822041      PMCID: PMC3638182          DOI: 10.1136/amiajnl-2012-000928

Source DB:  PubMed          Journal:  J Am Med Inform Assoc        ISSN: 1067-5027            Impact factor:   4.497


Introduction

The identification of primary and recurrent cancer diagnoses is critical to clinical and epidemiologic research. Despite its importance, however, there is substantial lag between the time of primary cancer diagnoses and complete information capture by cancer registries. Additionally, many registries do not track cancer recurrences. Consequently, researchers frequently rely on manual chart review and medical claims data, such as International Classification of Diseases (ICD) codes, to identify primary and recurrent cancers. Chart review, while accurate, is often not feasible in large-scale studies. Conversely, claims data are inexpensive and scalable to large studies but are unreliable in terms of accuracy. To address ongoing needs for improved identification of cancer diagnoses, a SAS-based coding, extraction, and nomenclature tool (SCENT) was developed at Kaiser Permanente Southern California (KPSC). KPSC is an integrated healthcare organization that provides medical services to a diverse membership of more than 3.5 million people throughout Southern California. Research conducted at KPSC directly impacts practice guidelines and the medical care that patients receive. Each year, approximately 20 000 new cancer cases are diagnosed at KPSC. Given the threat to patients' health and quality of life, as well as the impact on their families, prevention and treatment of cancer are high priorities for clinicians and researchers. SCENT was developed to support these critical efforts.

Background and significance

In 1999, Warren et al analyzed the medical claims data of 6784 Medicare patients and concluded that such data have limited value in accurately identifying breast cancer cases.1 More recently, Lamont et al achieved high levels of sensitivity and specificity using claims data to identify cancer recurrence.2 However, that validation used a small sample of patients (N=45) and the accuracy of claims data in identifying cancer recurrence has yet to be well established. The use of electronic medical record (EMR) systems among healthcare providers has increased significantly over the past decade. Additionally, a provision in the 2009 American Recovery and Reinvestment Act may accelerate the adoption of electronic systems by providing financial incentives for the meaningful use of technology in healthcare delivery.3 According to an annual survey by the Centers for Disease Control and Prevention, the use of EMRs by office-based physicians increased from 18% to 57% between 2001 and 2011.4 Use of EMR systems complete with all basic functionality also rose from 11% to 34% between 2006 and 2011. The transition from paper to EMRs not only reduces medical errors and improves communication between providers, but also increases the value and feasibility of medical informatics applications. Using text from EMRs, natural language processing (NLP) has the potential to supplement or replace manual chart review and electronic diagnosis codes in identifying primary and recurrent cancers. A number of studies have already demonstrated the utility of NLP in coding and extracting information from clinical text.5–7 However, few epidemiologic studies have employed the technology. Slow adoption cannot be entirely attributed to the technology's recentness, as working implementations have existed for nearly two decades.8 9 More likely, adoption has been limited by requirements for integration with clinical data systems, technical complexity, and habitual use of medical claims data. In 2011, Chapman et al noted that despite continued improvements in NLP performance, the technology is rarely employed in clinical research settings.10 This is attributed, in part, to a lack of focus on the end-user during development, specifically as it relates to implementation costs and customizability. Furthermore, NLP functionality in EMR and other clinical systems software currently provides limited value to researchers due to lack of customizability and inability to be readily shared between collaborating institutions. SCENT was developed to increase the attractiveness of NLP to clinical and epidemiologic researchers by reducing implementation barriers and ensuring accessibility in collaborative multi-site research.

Materials and methods

Overview

SCENT uses functionality from SAS Base (V.9.2, SAS Institute) and does not require the Text Miner add-on module. Components of SCENT consist primarily of SAS macro libraries and collections of Excel support files. An overview of the processes employed by SCENT can be seen in figure 1. Hierarchical classification rules are used by SCENT to analyze electronic pathology report text. Rules-based NLP approaches have been used effectively in previous studies11 12 and Informatics for Integrating Biology and the Bedside challenges.13–15 While SCENT has the flexibility to assign codes to electronic text using a number of different coding systems, SNOMED 3.x was initially selected due to its use by the CoPathPlus (V.3.2; Cerner DHT, Inc., Waltham, Massachusetts, USA) laboratory information system at KPSC. CoPathPlus features synoptic reporting, a process by which pathologists provide structured results using predefined cancer checklists.16 Results are assigned SNOMED 3.x codes by CoPathPlus, which can be reviewed and subsequently modified by reporting pathologists.
Figure 1

Process diagram for a SAS-based coding, extraction, and nomenclature tool (SCENT).

Process diagram for a SAS-based coding, extraction, and nomenclature tool (SCENT). For clinical concepts to be matched by SCENT, their keywords must be found, without negation, in proximate distance to each other. SCENT assigns SNOMED codes associated with matched concepts to examined pathology reports. Words within reports that contribute to concept matches are tagged with relevant information, such as disease extent and SNOMED code. SCENT examines text surrounding concept matches to differentiate clinical suspicions from conclusive findings and to determine disease extent (eg, non-invasive, invasive, or metastatic). Additional diagnostic information, including tumor stage and Gleason score, is also extracted.

Clinical concepts

SCENT relies on a dictionary of approximately 1000 clinical concepts and associated SNOMED 3.x codes related to morphology, anatomic site, and procedural type. The malignancy potential of each morphology concept was classified by up to four physicians with expert pathology or oncology knowledge. Dictionary concepts are tokenized into their component words and enhanced with regular expression logic to account for synonyms and plural words. For example, the ‘intraductal papillary adenocarcinoma’ concept is broken into three distinct words and ‘intraductal’ is replaced with ‘((intra)?duct(al)?)’ to match mentions of intraductal, ductal, or duct. In addition to regular expressions that are manually added to individual concepts, some transformations are uniformly applied to the dictionary, including partial reverse lemmatization to account for plural words. Exclusionary words are also included for some concepts to prevent matches from occurring in unrelated contexts. The ‘neck’ anatomic site concept, for instance, includes ‘bladder’ as an exclusionary word to prevent matching when ‘bladder neck’ is encountered.

Preprocessing

To prepare electronic pathology report text for analysis, SCENT performs a number of preprocessing tasks. In the pathology database at KPSC, report text spanning multiple records is often split at locations other than sentence boundaries. To facilitate matching of clinical concepts, SCENT uses approximately 70 regular expressions to reconstruct reports for analysis. In addition to recreating sentence boundaries, special characters are removed and common abbreviations replaced. Sample pathology report text, in original and preprocessed forms, can be seen in figure 2.
Figure 2

Sample pathology report text following preprocessing by a SAS-based coding, extraction, and nomenclature tool (SCENT).

Sample pathology report text following preprocessing by a SAS-based coding, extraction, and nomenclature tool (SCENT).

Matching

Using a dictionary of clinical concepts, SCENT examines preprocessed pathology report text for concept matches. In preparation, a clinical concept dictionary is read into memory and stored as a SAS hash object for future use. The text of each report is split by sentence and stored in arrays of character variables. Sentences are then tokenized and their component words stored in arrays. Additional hash objects are used to record intermediate information, such as match results of individual concept words, match positioning, and negation status. After preparing the necessary arrays and hash objects, SCENT sequentially examines the sentences of each pathology report. The clinical concept dictionary is processed separately for each new sentence and a search for the words of each concept is performed. Unmatched words are recorded to prevent unnecessary subsequent searches with concepts containing previously unmatched words. The text surrounding matches is examined for potential negation and the positions of non-negated matches are recorded. Dictionary concepts are considered fully matched if all words are found in proximate distance to each other. This distance is a token constant, defined as 10 words in the current study, and ignores prepositions, articles, and certain other words. SNOMED codes associated with matched concepts are assigned to the report from which they were matched. Surrounding text is examined to classify disease extent (eg, non-invasive, invasive, or metastatic). As shown in figure 3, words from reports that contribute to concept matches are tagged with relevant information, such as disease extent and SNOMED code. SCENT applies a single tag to multiple match words if they are adjacent or separated by only common prepositions and articles.
Figure 3

Sample pathology report text following preprocessing and code assignment by a SAS-based coding, extraction, and nomenclature tool (SCENT).

Sample pathology report text following preprocessing and code assignment by a SAS-based coding, extraction, and nomenclature tool (SCENT). Using codes from matched clinical concepts and hierarchical decision rules, SCENT classifies the overall status of reports as either benign, borderline, basaloid, or malignant. The disease extent of assigned morphology codes and anatomic sites are used to differentiate between new primary and recurrent malignancies. To facilitate the process, anatomic site classifications are consolidated into categories. For example, the sternum and clavicle sites are considered part of the bone category. Sites relating to regional disease spread, such as the neck and groin, are consolidated into the lymph nodes category. In the case of identified cancer metastases, SCENT attempts to determine both origin and metastatic sites.

Negation and uncertainty

SCENT uses a collection of regular expressions to investigate potential negation of matched concepts. This approach is similar to that of the NegEx algorithm, which has been used successfully in multiple studies.17–19 SCENT contains approximately 40 negation expressions, each assigned values limiting the acceptable distance between themselves and concept matches. For example, the expression ‘free (of|from)’ will negate concept matches appearing up to 10 words before or after the expression. In contrast, the negation word ‘without’ is valid up to only four words before concept matches and zero following them. The maximum number of words permitted between concept matches and negation expressions is modified by the presence of certain words, such as coordinating conjunctions (eg, ‘and’). To differentiate clinical suspicions from conclusive findings, SCENT examines text surrounding matches for uncertainty cues, such as ‘suspicious’ and ‘uncertain.’ This process shares some conceptual similarity with the implementation of speculation cues by Clark et al in their MITRE system.20 The scope of uncertainty by SCENT, however, is determined using fixed token distances rather than analysis of linguistic structure and statistical classifiers.

Validation

A validation study was conducted under institutional review board approval using EMR records of breast and prostate cancer patients at KPSC. Electronic pathology reports were selected for validation due to their availability and significance in diagnosing most cancers. To assess the accuracy of SCENT, its classifications of pathology reports were compared to those of experienced chart abstractors. Classifications based on codes assigned by CoPathPlus were also compared. Breast cancer cases were diagnosed from 2000 to 2007 with American Joint Committee on Cancer (AJCC) stage 0–III tumors, and had no prior history of cancer. All prostate cancer cases diagnosed from 2000 to 2005 were included irrespective of stage and previous cancer history. To address needs for improved recurrence identification in ongoing research at KPSC, validation focused on the period following cancer diagnosis and treatment. Of the 400 patients randomly selected for each cancer type, 206 breast and 186 prostate cancer patients had one or more pathology reports during the period beginning 6 months after diagnosis and subsequent to primary treatment(s) through the end of 2008. In total, 490 breast and 425 prostate cancer patient pathology reports were reviewed by two trained abstractors, one for each cancer group. Abstractor reviews of pathology reports served as the gold standard. Each report was classified according to malignancy status. In the case of malignant and suspicious findings, abstractors also recorded anatomic site and any mention of metastasis. Non-melanoma skin malignancies were considered benign by abstractors, SCENT, and classifications based on CoPathPlus. Independent reviewers made final classifications for the small number of cases about which abstractors were uncertain. Reviewers also adjudicated the few cases of discordant classifications between SCENT and abstractors. To reduce the time and cost of the validation process, we used SCENT to output pathology text, as can be seen in figure 4. Concepts relating to anatomic site were highlighted in green, malignancies in shades of blue according to disease extent, and suspicious findings in orange. Abstractors were instructed to fully review the text of each output report. To investigate the potential for bias stemming from SCENT highlights, a sample of pathology reports was independently reviewed by two abstractors. One set of reports contained highlighted text, the other did not. A total of 73 reports were reviewed from the first 30 breast cancer patients in the random sample. Abstractor classifications were found to be identical. Performance of SCENT and classifications based on CoPathPlus were assessed using standard evaluation metrics, including: sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), and Cohen's κ.
Figure 4

Sample chart review form used by abstractors to classify the pathology reports of breast and prostate cancer patients.

Sample chart review form used by abstractors to classify the pathology reports of breast and prostate cancer patients.

Results

As shown in table 1, SCENT was highly concordant with abstractor classifications, achieving κ values of 0.96 and 0.95 in the breast and prostate cancer groups, respectively. SCENT identified 51 of 54 new primary and 60 of 61 recurrent cancer cases across both groups, with only three false positives in 792 true benign cases. Agreement was moderate for classifications based on the codes assigned by CoPathPlus, with κ values of 0.72 and 0.65 in the breast and prostate cancer groups, respectively. To calculate overall performance metrics, results were consolidated into two categories: benign/suspicious and primary/recurrent cancer. Within both the breast and prostate cancer groups, SCENT reached levels of or above 94% for sensitivity, specificity, PPV, and NPV (table 2). Specificity for classifications based on CoPathPlus exceeded 98% in both groups. PPV was moderately high in both the breast (90%) and prostate (88%) cancer groups. Sensitivity for classifications based on CoPathPlus in detecting true positive malignancies was moderate in the breast (74%) and prostate (71%) cancer groups.
Table 1

Pathology report classifications of a SAS-based coding, extraction, and nomenclature tool (SCENT) and Cerner's CoPathPlus coding, as compared with abstractor review

BenignCancer recurrenceOther primary cancerSuspiciousκ Value
%N%N%N%N
Breast cancer (abstractor) (436) (32) (18) (4)
 SCENT
  Benign99.843525.010.96
  Cancer recurrence100.032
  Other primary cancer0.21100.01850.02
  Suspicious25.01
 CoPathPlus
  Benign97.242412.5422.2450.020.72
  Cancer recurrence0.7384.427
  Other primary cancer55.61025.01
  Suspicious2.193.1122.2425.01
Prostate cancer (abstractor) (356) (29) (36) (4)
 SCENT
  Benign99.43545.620.95
  Cancer recurrence96.6282.81
  Other primary cancer0.623.4191.733
  Suspicious100.04
 CoPathPlus
  Benign96.134210.3333.31250.020.65
  Cancer recurrence0.3172.4215.6225.01
  Other primary cancer1.1413.8452.819
  Suspicious2.593.418.3325.01

Includes the pathology reports of randomly sampled breast and prostate cancer patients from 6 months after diagnosis and subsequent to primary treatment. Contralateral breast cancers were considered to be recurrences.

Table 2

Validation metrics for a SAS-based coding, extraction, and nomenclature tool (SCENT) and Cerner's CoPathPlus coding, as compared with abstractor review of pathology reports

Sensitivity* Specificity* PPV* NPV*
Breast cancer
 SCENT1.00 (0.93 to 1.00)0.99 (0.98 to 1.00)0.94 (0.85 to 0.98)1.00 (0.99 to 1.00)
 CoPathPlus0.74 (0.60 to 0.84)0.99 (0.98 to 1.00)0.90 (0.77 to 0.96)0.97 (0.95 to 0.98)
Prostate cancer
 SCENT0.97 (0.89 to 0.99)0.99 (0.98 to 1.00)0.97 (0.89 to 0.99)0.99 (0.98 to 1.00)
 CoPathPlus0.71 (0.59 to 0.80)0.98 (0.96 to 0.99)0.88 (0.77 to 0.95)0.95 (0.92 to 0.97)

Includes the pathology reports of randomly sampled breast and prostate cancer patients from 6 months after diagnosis and subsequent to primary treatment.

Shown with Wilson's 95% CI.

NPV, negative predictive value; PPV, positive predictive value.

Pathology report classifications of a SAS-based coding, extraction, and nomenclature tool (SCENT) and Cerner's CoPathPlus coding, as compared with abstractor review Includes the pathology reports of randomly sampled breast and prostate cancer patients from 6 months after diagnosis and subsequent to primary treatment. Contralateral breast cancers were considered to be recurrences. Validation metrics for a SAS-based coding, extraction, and nomenclature tool (SCENT) and Cerner's CoPathPlus coding, as compared with abstractor review of pathology reports Includes the pathology reports of randomly sampled breast and prostate cancer patients from 6 months after diagnosis and subsequent to primary treatment. Shown with Wilson's 95% CI. NPV, negative predictive value; PPV, positive predictive value. A secondary assessment of SCENT was conducted to evaluate its ability to extract AJCC tumor staging and Gleason score information. SCENT identified and accurately extracted all tumor (T), lymph node (N), and metastasis (M) staging information from 19 pathology reports across both cancer groups. Within the prostate cancer group, SCENT identified all Gleason scores and accurately extracted the tumor pattern scores from each of the 20 reports. There were no instances in which SCENT failed to identify reports containing tumor staging or Gleason score information.

Discussion

The favorable validation results in this study suggest that SAS-based NLP can be used to accurately identify and extract information from electronic clinical text. Using the pathology reports of patients previously diagnosed and treated for breast and prostate cancer, SCENT successfully identified 51 of 54 primary and 60 of 61 recurrent cancers. Additionally, there were only three false positive identifications within 793 true benign results. Performance was similar for both the breast and prostate cancer groups, with measures of sensitivity, specificity, PPV, and NPV of or above 94%. Classifications using codes assigned by CoPathPlus were moderately successful in identifying incident and recurrent cancers within both cancer groups. Sensitivity in the breast and prostate cancer groups was 74% and 71%, respectively, while PPV was 90% and 88%. To confirm the accuracy of abstractor classifications, an oncologist reviewed pathology reports associated with discordant CoPathPlus cases. This review identified only one abstractor classification error, which was subsequently corrected. Codes assigned to discordant CoPathPlus cases were reviewed to better understand the sources of classification error. Primary sources of error for classifications based on CoPathPlus were: coding of historical malignancies, assignment of overly general codes, and lack of code assignment for unknown reasons. Additionally, there were three false positives relating to the coding of ‘residual’ malignancy, which had been excluded a priori from study definitions of primary and recurrent cancer. The widespread adoption of SAS in clinical analysis and research settings ensures that SCENT is highly accessible. Integration of SAS with relevant data systems has already been established in these settings, allowing electronic text to be readily extracted for analysis. By avoiding the need for additional software installation or systems integration, technical and administrative barriers to NLP implementation are reduced. Additionally, SCENT does not require annotated training data and can be used by staff without specialized informatics or machine learning knowledge. Primary and recurrent cancers are frequently among the main study outcomes in epidemiologic cancer research. Despite the importance of accurately identifying these events, electronic diagnosis codes often lack accuracy and cancer registry data are not always available. SCENT achieved high levels of sensitivity and specificity in identifying pathologically diagnosed malignancies among patients previously diagnosed with breast or prostate cancer. In addition to its utility in retrospective research, SCENT can be used in prospective research and population care management. Cancer registry data, when available, are commonly delayed for months after patient diagnosis. It is often impractical to rely on these data for intervention studies relating to cancer treatment. SCENT is currently being implemented in one such study aimed at improving adherence to adjuvant hormonal therapy among breast cancer patients. Beyond its value in identifying pathologically diagnosed primary and recurrent cancers, SCENT has the potential for numerous other clinical applications. Questions investigated by clinical analysts are often central to medical leadership and decisions relating to patient care. Analysts face issues similar to those of prospective research studies, relying on electronic diagnosis codes and available cancer registry data to assess the quality of cancer care. SCENT has the potential to enhance the availability and depth of data and could be used to identify appropriate patient populations for evaluation. For example, while treatment recommendations for certain precancerous lesions may dictate watchful waiting at present, SCENT could be used to identify affected patients if those recommendations were to change. Resource requirements for chart review of uncoded electronic text can vary greatly between studies. Extraction of even small numbers of basic data elements can be infeasible in large sample studies due to labor costs and time requirements. As described in the Materials and methods section and illustrated in figure 4, SCENT can be used to highlight desired clinical concepts within electronic text. This functionality has the potential to dramatically reduce the time and costs associated with chart review. The benefits of SCENT in expediting chart review will be further explored and quantified in future work. While SCENT performed well in classifying and extracting information from electronic pathology reports, its performance with other clinical text sources is currently untested. These sources will likely necessitate modifications to the preprocessing and decision rules employed by SCENT. Modifications to preprocessing rules may also be needed prior to implementing SCENT outside of KPSC. Additionally, text sources such clinical progress notes are less structured and have greater linguistic variability. Statistical NLP approaches are likely to provide superior performance with respect to these sources. SCENT does not use formal part-of-speech tagging and is therefore limited in its ability to disambiguate and contextualize identified clinical concepts. The performance of SCENT relative to that of a general purpose NLP solution will need to be assessed across multiple text sources to identify performance gaps and inform appropriate usage. Future development and validation efforts for SCENT will include identification of incident and recurrent cancer cases that are diagnosed without pathological testing. Radiology reports and clinical progress notes are expected to be the primary sources for identifying non-pathologically diagnosed cancer cases. These sources pose additional challenges due to their diagnostic uncertainty and linguistic variability. SCENT may also have generalized utility extracting information outside of the oncology domain, such as standard scoring of cognitive impairment from the Mini–Mental State Examination (MMSE) and depression from the Patient Health Questionnaire (PHQ-9). Functionality was developed to assign confidence scores to individual SCENT dictionary concepts according to their false-positive rates from chart review results. These scores can be used in conjunction with specified minimum confidence thresholds to fit the sensitivity requirements and false-positive tolerance of individual studies. This process will be updated to use Bayesian methods for incorporating additional evidence from subsequent chart reviews. The feasibility of adopting additional statistical methods to enhance the rules-based classification approach of SCENT will also be assessed.

Conclusion

SCENT was highly successful in identifying and extracting information on primary and recurrent cancers from electronic pathology reports. This functionality has the potential to provide significant value to clinical and epidemiologic researchers, particularly when statistical NLP is infeasible due to resource or other constraints. SCENT is proof of concept for SAS-based NLP applications that can be easily shared between institutions and used to support clinical and epidemiologic research.
  20 in total

1.  MITRE system for clinical assertion status classification.

Authors:  Cheryl Clark; John Aberdeen; Matt Coarr; David Tresner-Kirsch; Ben Wellner; Alexander Yeh; Lynette Hirschman
Journal:  J Am Med Inform Assoc       Date:  2011-04-22       Impact factor: 4.497

2.  A simple algorithm for identifying negated findings and diseases in discharge summaries.

Authors:  W W Chapman; W Bridewell; P Hanbury; G F Cooper; B G Buchanan
Journal:  J Biomed Inform       Date:  2001-10       Impact factor: 6.317

3.  Implementation and evaluation of a negation tagger in a pipeline-based system for information extract from pathology reports.

Authors:  Kevin J Mitchell; Michael J Becich; Jules J Berman; Wendy W Chapman; John Gilbertson; Dilip Gupta; James Harrison; Elizabeth Legowski; Rebecca S Crowley
Journal:  Stud Health Technol Inform       Date:  2004

4.  Health information technology: revisions to initial set of standards, implementation specifications, and certification criteria for electronic health record technology. Interim final rule with request for comments.

Authors: 
Journal:  Fed Regist       Date:  2010-10-13

5.  A rule-based approach for identifying obesity and its comorbidities in medical discharge summaries.

Authors:  Ninad K Mishra; David M Cummo; James J Arnzen; Jason Bonander
Journal:  J Am Med Inform Assoc       Date:  2009-04-23       Impact factor: 4.497

6.  Electronic health record systems and intent to apply for meaningful use incentives among office-based physician practices: United States, 2001-2011.

Authors:  Chun-Ju Hsiao; Esther Hing; Thomas C Socey; Bill Cai
Journal:  NCHS Data Brief       Date:  2011-11

7.  Overcoming barriers to NLP for clinical text: the role of shared tasks and the need for additional creative solutions.

Authors:  Wendy W Chapman; Prakash M Nadkarni; Lynette Hirschman; Leonard W D'Avolio; Guergana K Savova; Ozlem Uzuner
Journal:  J Am Med Inform Assoc       Date:  2011 Sep-Oct       Impact factor: 4.497

8.  Comparing methods for identifying pancreatic cancer patients using electronic data sources.

Authors:  Jeff Friedlin; Marc Overhage; Mohammed A Al-Haddad; Joshua A Waters; J Juan R Aguilar-Saavedra; Joe Kesterson; Max Schmidt
Journal:  AMIA Annu Symp Proc       Date:  2010-11-13

9.  Natural language processing for the development of a clinical registry: a validation study in intraductal papillary mucinous neoplasms.

Authors:  Mohammad A Al-Haddad; Jeff Friedlin; Joe Kesterson; Joshua A Waters; Juan R Aguilar-Saavedra; C Max Schmidt
Journal:  HPB (Oxford)       Date:  2010-12       Impact factor: 3.647

10.  Description of a rule-based system for the i2b2 challenge in natural language processing for clinical data.

Authors:  Lois C Childs; Robert Enelow; Lone Simonsen; Norris H Heintzelman; Kimberly M Kowalski; Robert J Taylor
Journal:  J Am Med Inform Assoc       Date:  2009-04-23       Impact factor: 4.497

View more
  30 in total

1.  Using natural language processing to extract mammographic findings.

Authors:  Hongyuan Gao; Erin J Aiello Bowles; David Carrell; Diana S M Buist
Journal:  J Biomed Inform       Date:  2015-02-03       Impact factor: 6.317

2.  Creating a Synthetic Clinical Trial: Comparative Effectiveness Analyses Using an Electronic Medical Record.

Authors:  Marjorie G Zauderer; Aleksandr Grigorenko; Paul May; Nicholas Kastango; Isaac Wagner; Aryeh Caroline; Mark G Kris
Journal:  JCO Clin Cancer Inform       Date:  2019-06

3.  Electronic health records-driven phenotyping: challenges, recent advances, and perspectives.

Authors:  Jyotishman Pathak; Abel N Kho; Joshua C Denny
Journal:  J Am Med Inform Assoc       Date:  2013-12       Impact factor: 4.497

4.  Contralateral Breast Cancer Event Detection Using Nature Language Processing.

Authors:  Zexian Zeng; Xiaoyu Li; Sasa Espino; Ankita Roy; Kristen Kitsch; Susan Clare; Seema Khan; Yuan Luo
Journal:  AMIA Annu Symp Proc       Date:  2018-04-16

5.  tbiExtractor: A framework for extracting traumatic brain injury common data elements from radiology reports.

Authors:  Margaret Mahan; Daniel Rafter; Hannah Casey; Marta Engelking; Tessneem Abdallah; Charles Truwit; Mark Oswood; Uzma Samadani
Journal:  PLoS One       Date:  2020-07-01       Impact factor: 3.240

6.  Limited validity of diagnosis codes in Medicare claims for identifying cancer metastases and inferring stage.

Authors:  Neetu Chawla; K Robin Yabroff; Angela Mariotto; Timothy S McNeel; Deborah Schrag; Joan L Warren
Journal:  Ann Epidemiol       Date:  2014-07-03       Impact factor: 3.797

7.  Electronic Health Record (EHR) Abstraction.

Authors:  Amal A Alzu'bi; Valerie J M Watzlaf; Patty Sheridan
Journal:  Perspect Health Inf Manag       Date:  2021-03-15

8.  Clinical Informatics Researcher's Desiderata for the Data Content of the Next Generation Electronic Health Record.

Authors:  Timothy I Kennell; James H Willig; James J Cimino
Journal:  Appl Clin Inform       Date:  2017-12-21       Impact factor: 2.342

9.  Identifying Cases of Metastatic Prostate Cancer Using Machine Learning on Electronic Health Records.

Authors:  Martin G Seneviratne; Juan M Banda; James D Brooks; Nigam H Shah; Tina M Hernandez-Boussard
Journal:  AMIA Annu Symp Proc       Date:  2018-12-05

10.  A Systematic Review of Estimating Breast Cancer Recurrence at the Population Level With Administrative Data.

Authors:  Hava Izci; Tim Tambuyzer; Krizia Tuand; Victoria Depoorter; Annouschka Laenen; Hans Wildiers; Ignace Vergote; Liesbet Van Eycken; Harlinde De Schutter; Freija Verdoodt; Patrick Neven
Journal:  J Natl Cancer Inst       Date:  2020-10-01       Impact factor: 13.506

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.