Literature DB >> 25848598

Automating data abstraction in a quality improvement platform for surgical and interventional procedures.

Meliha Yetisgen1, Prescott Klassen1, Peter Tarczy-Hornoch1.   

Abstract

OBJECTIVE: This paper describes a text processing system designed to automate the manual data abstraction process in a quality improvement (QI) program. The Surgical Care and Outcomes Assessment Program (SCOAP) is a clinician-led, statewide performance benchmarking QI platform for surgical and interventional procedures. The data elements abstracted as part of this program cover a wide range of clinical information from patient medical history to details of surgical interventions.
METHODS: Statistical and rule-based extractors were developed to automatically abstract data elements. A preprocessing pipeline was created to chunk free-text notes into its sections, sentences, and tokens. The information extracted in this preprocessing step was used by the statistical and rule-based extractors as features.
FINDINGS: Performance results for 25 extractors (14 statistical, 11 rule based) are presented. The average f1-scores for 11 rule-based extractors and 14 statistical extractors are 0.785 (min=0.576,max=0.931,std-dev=0.113) and 0.812 (min=0.571,max=0.993,std-dev=0.135) respectively. DISCUSSION: Our error analysis revealed that most extraction errors were due either to data imbalance in the data set or the way the gold standard had been created.
CONCLUSION: As future work, more experiments will be conducted with a more comprehensive data set from multiple institutions contributing to the QI project.

Entities:  

Year:  2014        PMID: 25848598      PMCID: PMC4371448          DOI: 10.13063/2327-9214.1114

Source DB:  PubMed          Journal:  EGEMS (Wash DC)        ISSN: 2327-9214


Introduction

The Surgical Care and Outcomes Assessment Program (SCOAP) is a collaboration of multiple hospitals in Washington state with the purpose of improving the quality and comparing the effectiveness of surgical procedures.1 The research is led by clinicians in 55 hospitals and covers multiple types of surgery. Contributions of the SCOAP program have resulted in reduced surgical complications and in cost savings. One of the barriers to scaling up SCOAP is the lack of automated data collection. Data from each patient is manually abstracted from unstructured free-text clinical notes and structured clinical records and is entered through a web-based form into a database. The data collection forms are complex and can include more than 700 individual data elements. The average time spent on manual abstraction for a given case is between 35–40 minutes depending on the complexity of the form and the details of the case. To address the costly manual abstraction approach, the Agency for Healthcare Research and Quality (AHRQ) funded the Comparative Effectiveness Research and Translation Network (CERTAIN) initiative, a project designed to strengthen SCOAP by adding automated data abstraction and to prove its utility to comparative effectiveness research. To achieve this, SCOAP CERTAIN investigators implemented a clinical data repository (CDR) in which data is automatically retrieved from original source hospitals and stored for later analysis. The investigators analyzed the data elements in the data collection forms to identify data elements that have the potential to be automatically abstracted from structured data and unstructured clinical records. They specifically targeted data elements in the general SCOAP form, which is used to abstract information about general surgical procedures (e.g., appendectomies, colectomies, and bariatric surgeries) and includes multiple data elements that range from simple demographic information such as age, gender, and insurance, to more complex data such as whether there was an unplanned intensive care unit (ICU) stay in the postoperative period. Figure 1 presents one section of the form that collects information about indication of operation.
Figure 1.

Indication for Operation Section of SCOAP General Data Form

A subset of the form is common to all types of procedures (core data elements) and includes information such as the procedure date and surgeon information. The rest of the form is specific to the details of each type of procedure, such as the presence or absence of a colostomy in a procedure involving the colon. In a prior study, investigators analyzed each data element and identified 64.1 percent of the data elements as being manually abstracted from free-text clinical notes.2 We built a text-processing pipeline based on natural language processing (NLP) and machine learning to automatically abstract a subset of data elements from clinical notes. The main focus of this paper is to describe the text processing pipeline and its performance in the automatic abstraction task for the 25 data elements selected from the four main sections of the general SCOAP form.

Background and Related Work

Most information regarding patient state, diagnostic procedure, and disease progress is described in the narrative sections and semistructured lists or free-text fields of patient clinical records. These free-text parts of clinical notes provide an opportunity for NLP technologies to play a major role in clinical care research and to facilitate the analysis of information that otherwise has only been accessible through manual chart abstraction. NLP methods have been used in a variety of health care applications including development of decision support tools, quality improvement, and automated encoding for clinical research.3–4 There are several general-purpose NLP systems that perform the task of taking free-text clinical records and transforming them into a set of structured data elements. One such system is MedLEE.5 Its preprocessing stage includes sentence-breaking, tokenization (to break text into words), and POS tagging modules. In its main processing module, sentences are parsed using cascades of hand-coded rules and are ultimately transformed into semantic frames relevant to the extraction task at hand. In a very recent study, MedLEE has been applied to identify comorbidities from admission notes.6 The HiTEX system7–8 provides a set of extensible modules that can be combined into text processing pipelines and includes a section splitter, section filter, tokenizer, Part-of-Speech (POS) tagger, noun phrase finder, Unified Medical Language System (UMLS) concept finder, negation finder, regular expression-based concept finder, and sentence splitter. Text analytics systems released by Open Health Natural Language Processing Consortium (OHNLPC) (http://www.ohnlp.org/index.php/Main_Page) are designed to identify clinically relevant named entities in clinical notes, which can then be used in information retrieval and data mining tasks. Two Apache UIMA-based (http://uima.apache.org) text analytics systems have been used to build OHNLPC compliant NLP pipelines: medKAT, which extracts cancer characteristics from pathology reports, and cTAKES9 which is more general purpose and has been applied to varied tasks including the identification of disorders, drugs, anatomical sites, and procedures in clinical notes. Our approach is designed to be a highly customizable lightweight NLP pipeline that enables the rapid prototyping of text classification tasks based on a simple set of XML-based templates and the integration of existing standalone NLP tools (openNLP, libSVM, and Mallet). We decided to build our own lightweight pipeline after a review of available generalized clinical systems, which for various reasons did not meet the specific requirements of our project. Some systems (MedLEE) are proprietary and others (cTAKES, HiTEX) rely on third party frameworks like Apache UIMA and GATE, which we felt required too much time and effort to extend and customize to accommodate our particular data and system requirements. We considered another freely available standalone system, MediClass10 but it did not include modules related directly to our task. After careful consideration, we decided to build our own simple, lightweight NLP pipeline, integrating existing standalone NLP and machine learning tools.

Methods

The main components of our text-processing pipeline are depicted in Figure 2. The data set and components of the pipeline are described in the following sections.
Figure 2.

SCOAP Text Processing Pipeline

Data Set

In order to develop and evaluate our text-processing pipeline, we created a data set consisting of free-text reports pulled from 618 general surgical encounters performed at University of Washington Medical Center and Harborview Medical Center in 2010. The retrospective review of reports was approved by the University of Washington Human Subjects Committee of Institutional Review Board. A total of 20,760 reports (averaging just over 33 reports per encounter) were selected for inclusion in the data set and represented a wide range of report types (e.g., admit note, operative note, nursing notes, radiology reports, discharge summary) created between admission and discharge times of the patient. The reports capture the details of the surgery and overall state of the patient. Annotators were trained in manual review of medical records for SCOAP data collection. They filled out the general form by abstracting data elements from 618 encounters contained in the clinical reports. We used the abstractions over patient encounters as a gold standard to train and test our pipeline. Because the annotation task was completed by the SCOAP team before we initiated the NLP project, each encounter was annotated by only a single annotator and interrater agreement could not be calculated. We call this data set 2010 SCOAP general data set.

Text Processing Pipeline

The input to our pipeline is a request consisting of two components: a data element and its unique identifier from the general SCOAP form and a set of free-text records documenting a patient encounter (Figure 2). Each note in the patient record encounter is sent to a preprocessing pipeline and the resulting metadata is cached. All records and metadata are then routed to the information extraction component and associated with the requested data element. If reports have already been preprocessed in a previous session, the preprocessing step is skipped, and the cached metadata for the reports is retrieved from the cache and associated with the requested data element. The output is the result of rule-based or statistical classification processed by the extraction component. In the following sections, we describe the main components of our pipeline.

Preprocessing

The preprocessing component is a customizable NLP pipeline that takes a medical record as input and produces for output a version of the medical record that has been annotated with metadata created in each step of the pipeline. The steps of the pipeline are implemented in five components including the following: (1) a tokenizer, (2) a sentence breaker, (3) a section chunker, (4) a concept tagger, and (5) a negation analyzer. The components are applied to each record sequentially. For example, the sentence breaker used by the pipeline requires that the input already be tokenized. Therefore, it requires that the tokenizer must be run before the sentence breaker. We use the OpenNLP English tokenizer and sentence breaker (http://opennlp.apache.org) to tokenize and split text into sentences. We then apply a statistical approach to section chunking based on our previous work identifying the boundaries and types of the sections in clinical records.11 Our approach requires (1) the construction of an ontology of section headers for a selected clinical report type (e.g., discharge summary), and (2) the annotation of a corpus of notes based on the constructed ontology. The annotated notes serve as a training corpus for our statistical section chunker. In our approach to section segmentation each line in a document is classified based on its probability of inclusion in a section type. We incorporated two separate models and steps for section segmentation and classification. First, for section segmentation, section boundaries are identified by labeling each line with a “B” (beginning of section), an “I” (inside of section), or an “O” (outside of section) tag. The same text used in the first step is then passed unlabeled to the second step for section classification, where a separate classifier is called upon to label each section with the appropriate section category. We achieved 0.921 f1-score in chunking radiology reports and 0.968 f1-score in chunking discharge summaries based on a gold standard composed of manually annotated (100) radiology reports and (191) discharge summaries with respect to both the section segmentation and section classification tasks. After chunking reports into sections, we use MetaMap12–13 to identify the UMLS concepts and NegEx14 to extract the negation state of the identified concepts.

Information Extraction

We developed both rule-based and statistical information-extraction approaches to extract 25 data elements from the 2010 SCOAP data collection form and general data set. A total of 25 extractors were created: 11 rule based and 14 statistical. The 25 data elements were selected based on the following two criteria: the interest value for the whole project team, and the size and distribution of the data set (not all data elements have coverage across all encounters in the data set).

Rule-Based Extraction

Rule-based extraction uses one or more text-based triggers to classify patient records with an appropriate label. There are two basic types of triggers: medical concepts; and keywords and regular expressions. Medical concepts are sourced from the UMLS ontology and each report is processed by MetaMap to identify the medical concepts. The manually selected UMLS concepts are compared against those extracted by MetaMap. Keywords and regular expressions are manually formulated to cover the cases where UMLS concepts are not sufficient. These two types of triggers can be used in combination to define an overall classification rule. The classifier accepts one of the following three types of decision-making meta rules to apply the text-based triggers: (1) an instance voting threshold, (2) an instance voting maximum, or (3) an aggregate classification. Figure 3 includes an example rule we wrote to capture hypertension.
Figure 3.

Rule Set for Hypertension

The decisions to use specific keywords and regular expressions or UMLS concepts are made based on a combination of expert knowledge and tests of the rules’ effectiveness when used against a set of development data. If a UMLS concept can be found for a data element, its effectiveness as a trigger is evaluated against the set of development data to determine if the addition of regular expressions or keywords can improve performance. A polarity flag determines if the rule-based trigger should be considered neutral, positive, or negative. By default, all triggers have neutral polarity. Negation is addressed by associating a regular expression or keyword trigger with a negative polarity flag. Rule-based triggers are not constrained to only binary classification tasks. A trigger can be explicitly associated with one of many labels to address multiclass classification tasks.

Statistical Extraction

We designed our statistical extraction approach as a text classification task of clinical text (e.g., operative note) into categories of the data element (e.g., Figure 1: data element—indication for operation; categories—appendectomy, bariatric surgery, colon). Table 1 includes the list of basic features used to represent the content of the medical text. Some of the data elements are described throughout the reports (e.g., full-text operative reports for operation of indication). On the other hand, some elements are described in only certain sections of the reports (e.g., smoking history). To address the second kind of data elements, in our preliminary experiments, we found that using only sentences in which the data elements were mentioned improved the classification performance. As an example, for smoking history, we process only discharge summaries, preanesthesia reports, and pain management reports. In this limited collection of reports, we first identify sentences (or 15 word windows) with mentions of a manually selected list of smoking-related concepts (e.g., packs), and represent the content with the combined mention-window text as a feature vector, consisting of unigram words.15
Table 1.

Features Used in Statistical Extraction

Feature TypeFeatures
Textuni-, bi-, trigrams
Knowledge-basedUMLS concepts
SemanticNegation
StructuralSection type
We use the Maximum Entropy16 implementation available in the Mallet toolkit (http://mallet.cs.umass.edu) as our algorithm for our all classification tasks.

Findings

We used 5-fold cross validation on the annotated data set for performance evaluation. In 5-fold cross validation, the original data set is randomly partitioned into five equal-size subsamples—where four subsamples are used as the training data to train a model and the remaining subsample is retained as the validation data to test the model. This cross validation process is repeated five times. Table 2 shows the 5-fold cross validation results achieved by our system on the 2010 SCOAP general data set for the 25 data elements. The results for each data element are reported in terms of precision, recall, f1-score (harmonic mean of precision and recall), and accuracy. We report performance values for each data element category (e.g., “yes,” “no”) since the importance of each category is equivalent for this study. The average f1-scores for 11 rule-based extractors and 14 statistical extractors are 0.785 (min=0.576,max=0.931,std-dev=0.113) and 0.812 (min=0.571,-max=0.993,std-dev=0.135) respectively. The performance of extractors across data elements varies. We analyzed the performance of extractors presented in Table 2 for each data element type in the following sections.
Table 2.

Performance Evaluation

Data Element Type#Data ElementExtraction ApproachCategoriesTPFPFNTNPreRecF1Acc
Comorbidities1HypertensionRule basedYes20026202160.8850.9090.8970.900
No21620262000.9150.8930.904
2DiabetesStatistical (Mention)Yes11413143210.8980.8910.8940.942
No32114131140.9580.9610.960
3AsthmaRule basedYes5513233710.8090.7050.7530.922
No3712313550.9420.9660.954
4Sleep apneaStatistical (Mention)Yes12211153140.9170.8910.9040.944
No31415111220.9540.9660.960
5CADStatistical (Mention)Yes165204210.7620.4440.5610.946
No421205160.9550.9880.971
6HIVRule basedYes21904410.0951.0000.1740.959
No44101921.0000.9590.979
Risk factors7Cigarette smokerStatistical (Mention)Yes3017483980.6380.3850.4800.868
No3984817300.8920.9590.925
Indication of operation8Op-indicationStatisticalcolon236313200.9870.9960.9920.993
appendectomy104024521.0000.9810.9900.996
bariatric surgery216113400.9950.9950.9950.996
9Cancer of colonRule basedYes157112040.6820.5770.6250.924
No204117150.9490.9670.958
10DiverticulitisRule basedYes251301990.6581.0000.7940.945
No199013251.0000.9390.968
11PolypsRule basedYes7652190.5380.5830.5600.954
No2195670.9780.9730.976
12Rectal prolapseRule basedYes4502280.4441.0000.6150.979
No2280541.0000.9790.989
13Bowel obstructionStatisticalYes7592160.5830.4380.5000.941
No2169570.9600.9770.969
14ColostomyStatisticalYes42102210.6670.2860.4000.949
No21110240.9570.9910.974
15Ulcerative colitisStatisticalYes22012141.0000.9570.9780.996
No21410220.9951.0000.998
16Crohn’s diseaseRule basedYes30801990.7891.0000.8820.966
No19908301.0000.9610.980
17StrictureRule basedYes182701920.4001.0000.5710.886
No192027181.0000.8770.934
Operation type (colon)18Right hemicolectomyStatisticalYes51061801.0000.8950.9440.975
No18060510.9681.0000.984
19Left hemicolectomyStatisticalYes17072131.0000.7080.8290.970
No21370170.9681.0000.984
20Low anterior resectionStatisticalYes5719291320.7500.6630.7040.797
No1322919570.8200.8740.846
21Abdominal perineal resectionStatisticalYes52112190.7140.3130.4350.945
No21911250.9520.9910.971
22Total abdominal colectomyRule basedYes234191910.8520.5480.6670.903
No191194230.9100.9790.943
23Stoma takedownStatistical (Mention)Yes20182171.0000.1000.1820.924
No21718020.9231.0000.960
24Perineal proctectomyRule basedYes52961970.1470.4550.2220.852
No19762950.9700.8720.918
25Abdominal protectomyStatistical (Mention)Yes177241890.7080.4150.5230.869
No189247170.8870.9640.924

Notes: TP: True positives, FP: False positives, FN: False negatives, FP: False positives, Pre: Precision, Rec: Recall, F1: F1-Score, Acc: Accuracy.

Comorbidities

In Table 2, comorbidities are automatically extracted from discharge summaries available in the data set. There are three rule-based and three statistical extractors trained to extract the comorbidities. Among the rule-based extractors, we achieve the best average f1-score for hypertension (Avg=0.900 where Yes: f1-score=0.897, No: f1-score=0.904) and the worst average f1-score for HIV (Avg=0.576 where Yes: f1-score=0.174, No: f1-score=0.979). The reason for the performance difference is that the data set is too skewed for HIV (Yes: 2, No: 460). Although we achieve perfect recall with the defined rules for the “Yes” HIV category, the precision is low (0.095) due to 19 false positive cases. For the statistical extractors, the average f1-scores for diabetes (Avg=0.925 where Yes: f1-score=0.894, No: f1-score=0.960) and sleep apnea (Avg=0.932 where Yes: f1-score=0.904, No: f1-score=0.960) are higher than those of coronary artery disease (CAD) (Avg=0.766 where Yes: f1-score=0.561, No: f1-score=0.971).

Risk Factors

We built a statistical extractor for smoking history, in Table 2. We processed discharge summaries, preanesthesia reports, and pain management reports in our extraction. The average f1-score is 0.868 (Yes: f1-score=0.480, No: f1-score=0.925).

Indication of Operation

Data elements from the “indication of operation” section of Table 2 are extracted from operative notes. The SCOAP general form data element for indication of operation covers three main surgery types including (1) colon, (2) appendectomy, and (3) bariatric surgery. We trained a three-way classifier to identify the surgery type for this data element. Operative reports provide an in-depth description of the procedure and, as a result, our classifier identifies the surgery type very accurately with f1-scores 0.992 for colon, 0.990 for appendectomy, 0.995 for bariatric surgery. The other nine data elements under this section are related to the details of the surgery performed (e.g., cancer of colon for colon surgery). We built rule-based extractors for six of the nine data elements (e.g., cancer of colon) and statistical extractors (e.g., bowel obstruction) for the remaining three data elements. Although, the performance is high for op-indication, for more detailed data elements, the performance is lower, especially for the cases where the data set was imbalanced (e.g., cancer of colon, bowel obstruction, colostomy). The average f1-scores for the rule-based and statistical classifiers are 0.821 and 0.803 respectively.

Operation Type (Colon)

Table 2’s “Operation type (colon)” section includes eight data elements that described the specific colon operation performed and were extracted from operative notes. We built two rule-based and six statistical extractors. The overall performance of the statistical classifiers is good in general with the exception of the classifier created for stoma takedown (Avg=0.571 where Yes: f1-score=0.182, No: f1-score=0.960). Out of the two rule-based extractors, the performance of the extractor for perineal proctectomy (Avg=0.570 where Yes: f1-score=0.222, No: f1-score=0.918) is lower than that of the extractor for abdominal perineal resection (Avg=0.703 where Yes: f1-score=0.435, No: f1-score=0.971).

Discussion

Our data set has limitations. First, the data set covers 618 patients from one institution. For some of the data elements (e.g., abdominal perineal resection), the data set is too imbalanced, which causes low performance for the underrepresented classes (Yes: 16, No: 221). Second, the annotation of the data set was completed before the initiation of the NLP project. The annotators had access to the complete patient charts while they manually abstracted the forms. While designing our system, we created—for each of the 25 data elements—a list of report types the annotators typically used for extraction. However, our error analysis revealed that many of the false negatives for a given data element were due to report types not processed by the NLP system for the given data element. In addition, some reports types are stored in the EMR as scans of handwritten documents (e.g., history and physical reports). We decided not to include such reports although we knew information about certain data elements (e.g., comorbidities) were extracted from those reports during manual abstraction. The performance results presented in Table 2 are lower bounds for the real system performance.

Conclusion

In this paper, we describe a text-processing pipeline based on statistical and rule-based approaches. We report performance results for 25 data elements collected for a surgical quality improvement program. We trained and tested our approaches on a limited data set composed of cases whose surgeries were performed in our institution during 2010. The overall performance of both statistical and rule-based extractors is encouraging. Our error analysis of the low performing extractors revealed that the size of the data set was not enough to capture the characteristics of some of the data elements. In addition, we achieved higher statistical extraction performance when the data sets were more balanced. The number of cases manually abstracted since we created the data set has increased dramatically. As future work, we plan to run experiments on a larger data set covering a longer period. We believe a larger data set will improve overall performance of the extractors. We also plan to extend our data set by including cases from other institutions contributing to SCOAP. This more comprehensive data set will enable us to run domain adaptability experiments to test the generalizability of our approaches. We released some components of the described text-processing pipeline including section chunker to the NLP community as open source tools. These tools are downloadable at our research website (http://depts.washington.edu/bionlp).
  12 in total

1.  Effective mapping of biomedical text to the UMLS Metathesaurus: the MetaMap program.

Authors:  A R Aronson
Journal:  Proc AMIA Symp       Date:  2001

2.  A simple algorithm for identifying negated findings and diseases in discharge summaries.

Authors:  W W Chapman; W Bridewell; P Hanbury; G F Cooper; B G Buchanan
Journal:  J Biomed Inform       Date:  2001-10       Impact factor: 6.317

3.  An overview of MetaMap: historical perspective and recent advances.

Authors:  Alan R Aronson; François-Michel Lang
Journal:  J Am Med Inform Assoc       Date:  2010 May-Jun       Impact factor: 4.497

4.  MediClass: A system for detecting and classifying encounter-based clinical events in any electronic medical record.

Authors:  Brian Hazlehurst; H Robert Frost; Dean F Sittig; Victor J Stevens
Journal:  J Am Med Inform Assoc       Date:  2005-05-19       Impact factor: 4.497

5.  A suite of natural language processing tools developed for the I2B2 project.

Authors:  Sergey Goryachev; Margarita Sordo; Qing T Zeng
Journal:  AMIA Annu Symp Proc       Date:  2006

6.  Deriving comorbidities from medical records using natural language processing.

Authors:  Hojjat Salmasian; Daniel E Freedberg; Carol Friedman
Journal:  J Am Med Inform Assoc       Date:  2013-10-31       Impact factor: 4.497

7.  Unlocking clinical data from narrative reports: a study of natural language processing.

Authors:  G Hripcsak; C Friedman; P O Alderson; W DuMouchel; S B Johnson; P D Clayton
Journal:  Ann Intern Med       Date:  1995-05-01       Impact factor: 25.391

Review 8.  What can natural language processing do for clinical decision support?

Authors:  Dina Demner-Fushman; Wendy W Chapman; Clement J McDonald
Journal:  J Biomed Inform       Date:  2009-08-13       Impact factor: 6.317

9.  Extracting principal diagnosis, co-morbidity and smoking status for asthma research: evaluation of a natural language processing system.

Authors:  Qing T Zeng; Sergey Goryachev; Scott Weiss; Margarita Sordo; Shawn N Murphy; Ross Lazarus
Journal:  BMC Med Inform Decis Mak       Date:  2006-07-26       Impact factor: 2.796

10.  Preparing Electronic Clinical Data for Quality Improvement and Comparative Effectiveness Research: The SCOAP CERTAIN Automation and Validation Project.

Authors:  Emily Beth Devine; Daniel Capurro; Erik van Eaton; Rafael Alfonso-Cristancho; Allison Devlin; N David Yanez; Meliha Yetisgen-Yildiz; David R Flum; Peter Tarczy-Hornoch
Journal:  EGEMS (Wash DC)       Date:  2013-09-10
View more
  6 in total

1.  Can Patient Record Summarization Support Quality Metric Abstraction?

Authors:  Rimma Pivovarov; Yael Judith Coppleson; Sharon Lipsky Gorman; David K Vawdrey; Noémie Elhadad
Journal:  AMIA Annu Symp Proc       Date:  2017-02-10

2.  Accelerating Chart Review Using Automated Methods on Electronic Health Record Data for Postoperative Complications.

Authors:  Zhen Hu; Genevieve B Melton; Nathan D Moeller; Elliot G Arsoniadis; Yan Wang; Mary R Kwaan; Eric H Jensen; Gyorgy J Simon
Journal:  AMIA Annu Symp Proc       Date:  2017-02-10

Review 3.  Aspiring to Unintended Consequences of Natural Language Processing: A Review of Recent Developments in Clinical and Consumer-Generated Text Processing.

Authors:  D Demner-Fushman; N Elhadad
Journal:  Yearb Med Inform       Date:  2016-11-10

4.  Applying Machine Learning Across Sites: External Validation of a Surgical Site Infection Detection Algorithm.

Authors:  Ying Zhu; Gyorgy J Simon; Elizabeth C Wick; Yumiko Abe-Jones; Nader Najafi; Adam Sheka; Roshan Tourani; Steven J Skube; Zhen Hu; Genevieve B Melton
Journal:  J Am Coll Surg       Date:  2021-04-05       Impact factor: 6.532

5.  Enhanced Quality Measurement Event Detection: An Application to Physician Reporting.

Authors:  Suzanne R Tamang; Tina Hernandez-Boussard; Elsie Gyang Ross; Gregory Gaskin; Manali I Patel; Nigam H Shah
Journal:  EGEMS (Wash DC)       Date:  2017-05-30

6.  Automating Electronic Clinical Data Capture for Quality Improvement and Research: The CERTAIN Validation Project of Real World Evidence.

Authors:  Emily Beth Devine; Erik Van Eaton; Megan E Zadworny; Rebecca Symons; Allison Devlin; David Yanez; Meliha Yetisgen; Katelyn R Keyloun; Daniel Capurro; Rafael Alfonso-Cristancho; David R Flum; Peter Tarczy-Hornoch
Journal:  EGEMS (Wash DC)       Date:  2018-05-22
  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.