Literature DB >> 31438083

Developing Customizable Cancer Information Extraction Modules for Pathology Reports Using CLAMP.

Ergin Soysal1, Jeremy L Warner2,3,4, Jingqi Wang1, Min Jiang1, Krysten Harvey3, Sandeep Kumar Jain5, Xiao Dong1, Hsing-Yi Song1, Harish Siddhanamatha1, Liwei Wang6, Qi Dai2, Qingxia Chen7, Xianglin Du8, Cui Tao1, Ping Yang6, Joshua Charles Denny2,3, Hongfang Liu6, Hua Xu1.   

Abstract

Natural language processing (NLP) technologies have been successfully applied to cancer research by enabling automated phenotypic information extraction from narratives in electronic health records (EHRs) such as pathology reports; however, developing customized NLP solutions requires substantial effort. To facilitate the adoption of NLP in cancer research, we have developed a set of customizable modules for extracting comprehensive types of cancer-related information in pathology reports (e.g., tumor size, tumor stage, and biomarkers), by leveraging the existing CLAMP system, which provides user-friendly interfaces for building customized NLP solutions for individual needs. Evaluation using annotated data at Vanderbilt University Medical Center showed that CLAMP-Cancer could extract diverse types of cancer information with good F-measures (0.80-0.98). We then applied CLAMP-Cancer to an information extraction task at Mayo Clinic and showed that we can quickly build a customized NLP system with comparable performance with an existing system at Mayo Clinic. CLAMP-Cancer is freely available for academic use.

Entities:  

Keywords:  Electronic Health Records; Information Storage and Retrieval; Natural Language Processing

Mesh:

Year:  2019        PMID: 31438083      PMCID: PMC7359882          DOI: 10.3233/SHTI190383

Source DB:  PubMed          Journal:  Stud Health Technol Inform        ISSN: 0926-9630


Introduction

Cancer is a highly heterogeneous group of related diseases with a growing body of evidence that cancer initiation, progression, metastasis, and response to treatment are directly associated with variability at the molecular level [1]. Technological advances have enabled us to identify an increasing number of molecular biomarkers, leading towards personalized oncology [2]. Personalized cancer research relies heavily on large clinical series of well annotated and high-quality data. One emerging research direction is to leverage large, real-world practice data stored in electronic health records (EHRs) to facilitate clinical and translational research [3]. However, much of this information (e.g., tumor characteristics) can only be found in narrative or semi-structured text in EHRs, such as pathology reports. Despite the move towards synoptic reporting by the pathology community [4], uptake of this formatting is not complete and synoptic reports are usually only issued for definitive resections, as mandated by the American College of Surgeon’s Commission on Cancer Standard 2.1. Natural language processing (NLP) technologies that can automatically extract and structure information from narrative documents have been extensively investigated in the medical domain and cancer sub-domain [5, 6]. In particular, many studies have focused on named entity recognition (NER), a fundamental NLP task that locates and classifies named entities to predefined categories (e.g., diseases, drugs, and lab tests). Several NLP systems have been developed to process pathology reports, which contain rich information about tumor specimen characteristics [7]. An earlier study by Xu et al. extended the existing MedLEE system to extract information from pathology reports to support cancer studies [8]. Crowley et al. reported the Cancer Tissue Information Extraction System (caTIES), which focuses on de-identification of cancer specimens for research purposes, along with information retrieval and concept coding functionalities [9]. MedKAT (MedTAS/P) is a rule-based system, which aims to extract cancer characteristics from pathology reports including anatomic site, histology, and grade [10]. Recently, Savova et al reported another rule-based system named DeepPhe, which can extract summary of cancer phenotypes including morphology, topology, procedure, and staging information from clinical documents [11]. Despite the success of reported use cases on cancer information extraction from clinical text, NLP has not been widely used for routine EHR-based cancer research, likely due to implementation barriers. Because of the diversity and complexity of clinical documentation, significant effort is required to develop custom clinical NLP systems or to extend existing systems for individual applications. In a study that aimed to locally adopt a smoking module in the cTAKES system [12], users had to write extensive code to re-train modules and re-define rules [13], which could be challenging for programmers without much NLP expertise. Additionally, completeness and the accuracy of the information remain as major obstacles in computational processing of clinical texts. CLAMP (Clinical Language Annotation, Modeling, and Processing) [14] is a general clinical NLP system that provides not only high-performance NLP components but also user-friendly interfaces for building customized NLP pipelines for individual needs. In this study, we describe our work on developing a library of information extraction modules for important tumor informiton in pathology reports within CLAMP. Using these built-in modules and user-friendly interfaces of CLAMP, users can quickly build customized NLP solutions to extract cancer information from their local pathology reports, with minimal programming effort. We believe such a tool would greatly benefit cancer researchers to leverage clinical text for their studies.

Methods

CLAMP follows the pipeline-based architecture as defined by the Apache UIMA™ (Unstructured Information Management Applications) framework, where an NLP system consists of multiple components in a specific order. The development of CLAMP Cancer Modules involved building cancer-specific NLP components and extending CLAMP interface that allows end users to customize and assemble different components into an NLP pipeline for their specific needs.

Developing Cancer-Specific NLP components

As the primary goal is to extract important information related to cancer, we limited the scope to the “diagnosis” section of pathology reports and followed these steps to build CLAMP Cancer Modules:

Define an Information Model for Pathology Reports

We reviewed available models including the Cancer Disease Knowledge Representation Model (CDKR) used in the MedKAT [10] and the information model in caTIES [9]. As our goal is to support clinical research, we specified data elements for pathology reports as recommended by the College of American Pathologists (CAP) (Figure 1). We reviewed all available CAP cancer templates (available at http://www.cap.org/cancerprotocols) as of January 2017, including the biomarker templates, and defined a list of minimum elements and relations to be included in CLAMP-Cancer Modules.
Figure 1.

CLAMP information model is based on data element suggestions of College of American Pathologists

We define Primary-site (site of the source of the material extracted from the body) as the root concept, with relations to Specimen (often numbered in the reports), Sub-site (detailed sub-anatomical sites related to the primary sites), Procedure (methods for specimen extraction from the body, e.g., biopsy), and Histology diagnosis (the morphology of cancer cells, e.g., “squamous cell carcinoma”). In addition, we also extract several important attributes defining histologic behavior and malignant potential, including Tumor Grade, Tumor Size, Invasiveness (invasion status and sites infiltrated by the tumor), Tumor Margin (tumor involvement in excisional margins), as well as Tumor Biomarkers and Values (e.g., estrogen and progesterone receptors).

Annotate a Corpus of Pathology Reports following the Information Model

Based on the information model, we developed an annotation guideline, which specifies the types of entities and relations for annotation, as well as different examples. We randomly selected 400 pathology reports from patients with a cancer ICD9 code at Vanderbilt University Medical Center (VUMC) in 2010. Four domain experts manually reviewed each pathology report and annotated all the entities and relations following the guideline, using CLAMP. Following a typical evaluation design for NLP studies, the annotated corpus was then divided into a training set (200 reports) and a test set (200 reports).

Develop Various Entity and Relation Extraction Components of CLAMP Cancer Modules

As CLAMP Cancer Modules extract a broad range of entities and their relations from pathology reports, different approaches were implemented to achieve optimal performance. For named entity recognition, we implemented regular expression-based, dictionary lookup-based, as well as machine learning-based approaches. For relation extraction, both rule-based and machine learning approaches were developed. Hybrid approaches that combine rules and machine learning were also used to optimize the performance. All the resources collected for developing these components (e.g., dictionaries, rules, and models) are available in CLAMP Cancer Modules and users can use them as-is or customize them to improve performance on their local data.

Evaluating CLAMP Cancer Modules

The evaluation of CLAMP Cancer Modules has two aspects. First, we developed the default CLAMP Cancer Modules using the training set from VUMC and evaluated them using the test dataset from VUMC. For each type of entity, we report precision, recall, and F-measure on recognizing entities only, using both exact and partial (relaxed) matching criteria [15]. We also report the system performance by considering both entity and relation information (e.g., a detected “sub_site” entity is correct only if both the entity and its relation to the linked “primary_site” are correct). Then, we apply CLAMP Cancer Modules to the same information extraction task as the MedKAT system [10]. Using CLAMP Cancer Modules, we assembled a customized NLP pipeline and compared it with MedKAT, using the same annotated corpus from Mayo Clinic.

Results

The developed CLAMP Cancer Modules consist of several components with machine learning models, dictionaries and rule sets to extract various types of cancer specific information from pathology reports following the CAP recomendations. Figure 2-a shows the cancer-specific modules in CLAMP and how they can be used to build customized NLP pipelines for local pathology reports. In Figure 2-b shows document annotation for cancer specific entities, and relationships among them.
Figure 2.

a. CLAMP Cancer module implements several cancer specific components (A), can be used to build customized pipelines (B). b. Pipeline annotates pathology reports to extract entities as well as their relationships.

Table 1 (a) shows the evaluation results of CLAMP Cancer Modules for different types of entities and relations. The default components of CLAMP Cancer Modules achieved good F-measures for entity recognition, with a range of 0.87 to 0.99, for exact matching criterion. A quick manual review shows that a fair number of relevant entities are missed by the system (e.g., a low recall in “invasion”), probably due to the complexity of expression patterns of these entities and the relatively limited size of the training corpus. When the relaxed matching criterion was used, CLAMP Cancer Modules achieved higher performance for some entities (e.g., Histology improved from 0.88 to 0.97), indicating some errors were about entity boundary only. When the relationships of entities were taken into consideration, performance dropped as expected (F-measures within 0.82-0.98), because of some known challenges for relation extraction (e.g., long-distance dependency). Table 1 (b) shows the results of the customized NLP pipeline built for Mayo Clinic’s pathology reports using CLAMP Cancer Modules, as well as the results from MedKAT. It took 12 hours for a developer to build the customized NLP pipeline for the specified task using CLAMP Cancer Modules, and it achieved comparable performance as MedKAT, indicating high effectiveness of CLAMP-Cancer for building customized NLP pipelines for pathology reports. Although MedKAT has slightly higher performance in certain entities, the production time and the adaptability remain as the major advantage of using CLAMP.
Table 1 –

Evaluation results of CLAMP Cancer Modules

Type of information# of ent.*Entity only (Exact/Relaxed Matching)Entity and Relation
PrecisionRecallF-measurePrecisionRecallF-measure
Specimen3100.99/0.990.99/0.990.99/0.990.970.980.98
Primary-site3510.98/0.990.98/0.990.98/0.990.980.980.98
Sub-Site1870.96/0.980.82/0.830.89/0.900.880.780.83
Procedure3390.98/0.990.98/0.990.98/0.990.970.970.97
Histology5530.91/1.000.85/0.930.88/0.970.900.850.86
Tumor Grade920.96/1.000.88/0.910.92/0.960.910.830.86
Tumor Size600.96/0.960.90/0.900.93/0.930.880.830.85
Tumor Margin930.92/0.990.91/0.980.92/0.980.800.790.80
Invasion710.92/1.000.83/0.900.87/0.950.860.780.82
Biomarker1070.95/0.990.90/0.940.92/0.960.880.840.86
(a) Performance of CLAMP-Cancer components on the VUMC test corpus* The number of each type of entities in the test corpus of 200 notes.
 
CLAMP-CancerMedKAT
PrecisionRecallF-1PrecisionRecallF-measure
Tumor size1.000.990.991.001.001.00
Dimension Extend1.000.990.990.991.000.99
Dimension Unit1.001.001.001.001.001.00
Tumor Site0.940.890.920.960.950.96
Histology0.910.920.920.960.980.97
Grade1.000.880.940.930.970.99
Date1.001.001.001.001.001.00
(b) Performance comparison of CLAMP Cancer Modules and MedKAT on the same information extraction task from 302 pathology reports at Mayo Clinic

The number of each type of entities in the test corpus of 200 notes.

Discussion and Conclusions

The contribution of CLAMP Cancer Modules is two-fold. First, it is a new NLP system that extracts comprehensive cancer information from pathology reports, including some that are not included by previous systems, e.g., biomarkers. Moreover, CLAMP Cancer Modules allow users to quickly build customized NLP pipelines by leveraging existing components and user-friendly interfaces. To the best of our knowledge, this is the first attempt to build cancer-focused NLP system that integrates into a drag and drop graphical user interface (GUI) for building customized NLP solutions for individualized needs. As pathology reports contain important cancer information, we believe that CLAMP Cancer will facilitate wide adoption of NLP in cancer research. We acknowledge that cancer knowledge is constantly evolving and new standards, such as the AJCC 8th Edition, may require re-annotation for additional concepts. As a full-featured NLP system, CLAMP Cancer Modules ease such tasks. To further accelerate EHR-based cancer research, we plan to extend CLAMP Cancer Modules in several directions. First, we will develop an encoding function to further map extracted cancer information to existing terminologies, e.g., ICD-O-3, or to standard value sets. We will then expand CLAMP Cancer Modules to include other types of clinical reports containing rich cancer information, e.g., radiology reports. We also plan to conduct formal usability studies of CLAMP Cancer Modules and iteratively improve the user interface based on users’ feedback.
  14 in total

1.  caTIES: a grid based system for coding and retrieval of surgical pathology reports and tissue specimens in support of translational research.

Authors:  Rebecca S Crowley; Melissa Castine; Kevin Mitchell; Girish Chavan; Tara McSherry; Michael Feldman
Journal:  J Am Med Inform Assoc       Date:  2010 May-Jun       Impact factor: 4.497

2.  2010 i2b2/VA challenge on concepts, assertions, and relations in clinical text.

Authors:  Özlem Uzuner; Brett R South; Shuying Shen; Scott L DuVall
Journal:  J Am Med Inform Assoc       Date:  2011-06-16       Impact factor: 4.497

3.  A study of transportability of an existing smoking status detection module across institutions.

Authors:  Mei Liu; Anushi Shah; Min Jiang; Neeraja B Peterson; Qi Dai; Melinda C Aldrich; Qingxia Chen; Erica A Bowton; Hongfang Liu; Joshua C Denny; Hua Xu
Journal:  AMIA Annu Symp Proc       Date:  2012-11-03

Review 4.  Electronic health records (EHRs): supporting ASCO's vision of cancer care.

Authors:  Peter Yu; David Artz; Jeremy Warner
Journal:  Am Soc Clin Oncol Educ Book       Date:  2014

Review 5.  Natural Language Processing in Oncology: A Review.

Authors:  Wen-Wai Yim; Meliha Yetisgen; William P Harris; Sharon W Kwan
Journal:  JAMA Oncol       Date:  2016-06-01       Impact factor: 31.777

Review 6.  Biomarkers for personalized oncology: recent advances and future challenges.

Authors:  Madhu Kalia
Journal:  Metabolism       Date:  2014-10-30       Impact factor: 8.694

7.  Cancer biomarker discovery and validation.

Authors:  Nicolas Goossens; Shigeki Nakagawa; Xiaochen Sun; Yujin Hoshida
Journal:  Transl Cancer Res       Date:  2015-06       Impact factor: 1.241

8.  Facilitating cancer research using natural language processing of pathology reports.

Authors:  Hua Xu; Kristin Anderson; Victor R Grann; Carol Friedman
Journal:  Stud Health Technol Inform       Date:  2004

9.  Standardized synoptic cancer pathology reporting: a population-based approach.

Authors:  John R Srigley; Tom McGowan; Andrea Maclean; Marilyn Raby; Jillian Ross; Sarah Kramer; Carol Sawka
Journal:  J Surg Oncol       Date:  2009-06-15       Impact factor: 3.454

10.  CLAMP - a toolkit for efficiently building customized clinical natural language processing pipelines.

Authors:  Ergin Soysal; Jingqi Wang; Min Jiang; Yonghui Wu; Serguei Pakhomov; Hongfang Liu; Hua Xu
Journal:  J Am Med Inform Assoc       Date:  2018-03-01       Impact factor: 4.497

View more
  3 in total

1.  Looking for low vision: Predicting visual prognosis by fusing structured and free-text data from electronic health records.

Authors:  Haiwen Gui; Benjamin Tseng; Wendeng Hu; Sophia Y Wang
Journal:  Int J Med Inform       Date:  2021-12-30       Impact factor: 4.046

2.  Automatic Classification of Cancer Pathology Reports: A Systematic Review.

Authors:  Thiago Santos; Amara Tariq; Judy Wawira Gichoya; Hari Trivedi; Imon Banerjee
Journal:  J Pathol Inform       Date:  2022-01-20

Review 3.  Assessment of Electronic Health Record for Cancer Research and Patient Care Through a Scoping Review of Cancer Natural Language Processing.

Authors:  Liwei Wang; Sunyang Fu; Andrew Wen; Xiaoyang Ruan; Huan He; Sijia Liu; Sungrim Moon; Michelle Mai; Irbaz B Riaz; Nan Wang; Ping Yang; Hua Xu; Jeremy L Warner; Hongfang Liu
Journal:  JCO Clin Cancer Inform       Date:  2022-07
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.