Literature DB >> 30474078

Learning relevance models for patient cohort retrieval.

Travis R Goodwin1, Sanda M Harabagiu1.   

Abstract

OBJECTIVE: We explored how judgements provided by physicians can be used to learn relevance models that enhance the quality of patient cohorts retrieved from Electronic Health Records (EHRs) collections.
METHODS: A very large number of features were extracted from patient cohort descriptions as well as EHR collections. The features were used to investigate retrieving (1) neurology-specific patient cohorts from the de-identified Temple University Hospital electroencephalography (EEG) Corpus as well as (2) the more general cohorts evaluated in the TREC Medical Records Track (TRECMed) from the de-identified hospital records provided by the University of Pittsburgh Medical Center. The features informed a learning relevance model (LRM) that took advantage of relevance judgements provided by physicians. The LRM implements a pairwise learning-to-rank framework, which enables our learning patient cohort retrieval (L-PCR) system to learn from physicians' feedback. RESULTS AND DISCUSSION: We evaluated the L-PCR system against state-of-the-art traditional patient cohort retrieval systems, and observed a 27% improvement when operating on EEGs and a 53% improvement when operating on TRECMed EHRs, showing the promise of the L-PCR system. We also performed extensive feature analyses to reveal the most effective strategies for representing cohort descriptions as queries, encoding EHRs, and measuring cohort relevance.
CONCLUSION: The L-PCR system has significant promise for reliably retrieving patient cohorts from EHRs in multiple settings when trained with relevance judgments. When provided with additional cohort descriptions, the L-PCR system will continue to learn, thus offering a potential solution to the performance barriers of current cohort retrieval systems.

Entities:  

Keywords:  information storage and retrieval; machine learning; medical informatics; search engine

Year:  2018        PMID: 30474078      PMCID: PMC6241510          DOI: 10.1093/jamiaopen/ooy010

Source DB:  PubMed          Journal:  JAMIA Open        ISSN: 2574-2531


OBJECTIVE

Electroencephalography (EEG) records the electrical activity along the scalp and measures spontaneous electrical activity of the brain, which makes it a primary tool for diagnosis of brain-related illnesses., But, as noted in Beniczky et al., the EEG signal is complex, and moreover, when EEG reports are created, the inter-observer agreement in EEG interpretation is known to be moderate. Both these problems can be addressed by providing clinical experts with the ability to automatically retrieve similar EEG signals and EEG reports through a patient cohort retrieval (PCR) system operating on an Electronic Health Record (EHR). A multi-modal EEG PCR system called MERCuRY was presented in Goodwin and Harabagiu, which leverages the heterogeneous nature of EEG data by processing both the clinical narratives from EEG reports as well as the raw electrode potentials derived from the recorded EEG signal data. Because the patient cohort criteria are expressed in natural language, the MERCuRY system is driven by its ability to rank relevant patients based on the narratives available from the EEG reports. However, as reported in Edinger et al. the current state-of-the-art methods are not yet satisfactory for retrieving relevant patients from clinical narratives. The primary objective of our study is the design, implementation and validation of a novel PCR system that learns how to optimally rank patients based on relevance judgements, providing improvements to current state-of-the-art methods. We demonstrate that by using a learning-to-rank framework informed by (1) features automatically extracted from the cohort description and from the clinical narratives and (2) physician feedback, we enhance the relevance of patients in cohorts by 27–53% above state-of-the-art. Moreover, our system for learning to rank patient cohorts is easily portable across EHR collections. Thus, our approach provides a framework for improving a PCR system when relevance judgements become available.

BACKGROUND AND SIGNIFICANCE

The automatic identification of patient cohorts satisfying a wide range of criteria—including clinical, demographic, and social information—has numerous applications, e.g. (1) clinical trial recruitment; (2) outcome prediction; and (3) survival analysis. Because patient cohort identification relies on the processing of EHRs, many systems use statistical techniques or machine learning methods informed by natural language processing of the clinical narratives. However, these systems cannot rank the identified patients based on the relevance to the cohort criteria. Relevance is at the core of information retrieval (IR) systems. Thus, viewing the problem of patient cohort identification as an IR problem, i.e. considering the problem of PCR, enables not only the identification of patients from a cohort, but to also the ranking of patients based on relevance to the inclusion and exclusion criteria used in the cohort description. Without considering patients’ relevance, patient cohort identification systems produce only a binary decision: the patient either belongs or does not belong to the cohort. Ranking of the patients in the cohort was essential in the usability studies performed with the MERCuRY system, as it enabled neurologist researchers to rapidly identify effective interventions for epilepsy accompanied by mental health comorbidities. However, not all the patients from the cohorts discovered by MERCuRY were relevant to the cohort criteria. Relevance judgements produced by neurologists indicated limitations of the system, but also provided important lessons that can be used for learning how to rank patients. Similarly, the analysis reported in Edinger et al. indicates the limitations of PCR systems developed for the TREC Medical Records (TRECMed) track in the annual Text REtrieval Conference (TREC) hosted by the National Institute for Standards and Technology (NIST). In the 2011 TRECMed evaluation, 24 PCR systems were tested against the same medical records and the same cohort descriptions, and their results were evaluated by 25 physicians. The relevance judgements produced during the TRECMed evaluations could also be used for learning how to rank patients. The MERCuRY system and most of the PCR systems participating in TRECMed had architectures similar to the one illustrated in Figure 1, providing a unifying framework for applying learning-to-rank. Learning-to-rank is a framework for using machine learning techniques for generating ranking models in IR. In the architecture illustrated in Figure 1, the cohort description was processed with the goal of generating a machine-readable query (an essential component of any IR system). Many PCR systems generated a query by relying on MetaMap, to discern concepts from the Unified Medical Language System (UMLS) from the cohort description; while some systems also used NegEx for detecting negated concepts. In addition, there were systems that (1) mapped the cohort description to ICD-9 codes or (2) simply considered bags of words. After the queries were constructed, systems expanded the query by introducing new terms (e.g. synonyms, related words, etc.) using different techniques, including (1) pseudo-relevance feedback (PRF); (2) applying Personalized PageRank to the UMLS Metathesaurus; (3) using semantic vectors provided by Random Indexing., Thus, each cohort description was transformed in a query through a query construction method and a query expansion method . Moreover, as illustrated in Figure 1, the EHR collection was indexed using Apache Lucene, Indri, or Terrier. The resultant index informed a variety of relevance models that produced an ordered list of medical records, . Because TRECMed focused on ranking hospital visits—groups of medical records generated during a patient’s stay—rather than individual medical records, teams considered methods to aggregate the relevance scores or rankings of retrieved medical records to produce a ranking of hospital visits, , informed by a report-to-visit mapping provided to TRECMed participants. The resulting visit ranking was, in some cases, filtered to account for specific cohort criteria such as age or gender, and/or re-ranked to produce the final ranking of visits, . The learning PCR system presented in this paper allows the various techniques used in the components of the TRECMed systems to be unified within a single architecture, enabling the exploration of the impact of each technique on the optimal ranking of patients.
Figure 1.

Architecture of a typical patient cohort retrieval system evaluated in TRECMed.

Architecture of a typical patient cohort retrieval system evaluated in TRECMed.

MATERIALS AND METHODS

Datasets and experimental settings

In this work, we explored the design of learning PCR systems in 2 settings: (1) a neurology-specific setting focusing on cohorts identified from a large archive of EEG reports and (2) a general setting allowing the recognition of cohorts from multiple forms of hospital records. Approval was obtained for both EHR collections from the Institutional Review Board (IRB) at the University of Texas at Dallas (UTD).

The Temple University electroencephalogram corpus

For the neurology-specific setting, we relied on the publicly-available collection of EEG reports from the Temple University Hospital (TUH) EEG Corpus., It contains EEG reports collected over 25 000 sessions for 15 000 patients over 12 years. While the TUH EEG corpus contains EEG signal information as well as EEG reports, we considered only the EEG reports.

The University of Pittsburgh

PCR systems participating in the TRECMed challenges had access to a large repository of 95 702 de-identified narratives from medical reports provided by the University of Pittsburgh Medical Center. This EHR repository consisted of 1 month of reports from multiple hospitals. Each medical report was associated with exactly 1 hospital visit (an individual patient’s single stay at a hospital). The data set contained 93 551 medical reports mapped into 17 264 visits.

Cohort descriptions

Patient cohorts were recognized from each of the EHR collections based on descriptions provided by practicing clinicians. When using the TUH EEG corpus, 30 cohort descriptions were generated by 4 practicing neurologists. For TRECMed, we used the official cohort descriptions released by the task organizers. Thirty-four cohort descriptions were evaluated in 2011 and 47 additional descriptions were evaluated in 2012. Examples of cohort descriptions used to train and evaluate the learning patient cohort retrieval (L-PCR) system are shown in Table 1.
Table 1.

Examples of cohort descriptions used to train and evaluate the learning cohort retrieval system

EEG reportsTRECMed 2011TRECMed 2012
Patients experiencing seizures and generalized shakingPatients with complicated GERD who receive endoscopyAdult patients with Alzheimer’s disease admitted from nursing homes with pressure ulcers
Multiple sclerosis and seizureWomen with osteopeniaElderly patients with subdural hematoma
Patients under 18-year-old with absence seizuresFemale patient with breast cancer with mastectomies during admissionPatients admitted with Hepatitis C and IV drug use
Patients over age 18 with history of developmental delay and EEG with electrographic seizuresAdult patients who are admitted with asthma exacerbationPatients treated for post-partum problems including depression, hypercoagulability, or cardiomyopathy
Patients evaluated for seizures vs strokePatients with CAD who presented to the Emergency Department with Acute Coronary Syndrome and were given PlavixPatients with inflammatory disorders receiving TNF-inhibitor treatment
Brain tumor and sharp waves, spike/polyspike, and wave or spikesChildren admitted with cerebral palsy who received physical therapyAdults under age 60 undergoing alcohol withdrawal
EEG showing triphasic wavesPatients co-infected with hepatitis C and HIVPatients with AIDS who develop pancytopenia
Patients with anoxic brain injury and EEG reports denoting brain deathAdult patients who presented to the emergency room with anion gap acidosis secondary to insulin dependent diabetesPatients with hypertension on anti-hypertensive medication
EEGs without sharp waves, spikes, or spike/polyspike and wave activity in patient’s diagnosed with epilepsyPatients with dementiaPatients taking atypical antipsychotics without a diagnosis schizophrenia or bipolar depression
EEG showing generalized periodic epileptiform dischargesCancer patients with liver metastasis treated in the hospital who underwent a procedurePatients who develop thrombocytopenia in pregnancy
Examples of cohort descriptions used to train and evaluate the learning cohort retrieval system

Relevance judgments

To train and evaluate the retrieval performance of our system, we used visit-level relevance judgments produced by neurologists and clinicians for each cohort description described above. In both collections, physicians were asked to judge visits retrieved for each cohort as being relevant, partiallyrelevant, or notrelevant to the cohort. Supplementary Material Appendix H provides details on how the judgments were obtained for each collection.

The learning patient cohort retrieval system

In this study, we focused on the design of a L-PCR system. Unlike traditional PCR systems, such as MERCuRY or those developed for TRECMed,, the L-PCR system uses a learning-to-rank approach for identifying patient cohorts that takes advantage of physician feedback. The learning-to-rank paradigm allows the L-PCR system to consider relevance judgments performed by clinicians to learn an improved patient relevance model used for retrieving and ranking patients for any given cohort descriptions. The L-PCR system illustrated in Figure 2 includes 5 main components:
Figure 2.

Architecture of the learning patient cohort retrieval system.

Architecture of the learning patient cohort retrieval system. a query processing component processes a given cohort description to produce a machine-readable query, ; an EHR processing component produces an index of the narratives from the EHR collection; a visit retrieval component retrieves a sub-set of “candidate” visits from the EHR collection, , to be ranked by the learning relevance model (LRM); a feature extraction component extracts features vectors corresponding to each candidate visit the relationship between the visit and the cohort description; and the learning relevance model uses a random forest (RF) classifier to infer the relevance scores for each candidate visit based on their associated feature vectors ; the RF is trained using the relevance judgments provided by physicians. An overview of each of these components is provided below, with additional details provided in Supplementary Material Appendices A–F.

Query processing

As with the typical PCR system illustrated in Figure 1, each cohort description is first processed by a query construction step followed by query expansion. Query construction. The L-PCR system incorporates 8 query construction methods, illustrated in Figure 3(a). Methods , , and represent the cohort description as a set of medical concepts detected by MetaMap, classified by a support vector machine (SVM),, or corresponding to the titles of Wikipedia articles,, respectively. By contrast, method represents the cohort description as a set (i.e. “bag”) of words. To account for the possibility of exclusion criteria in cohort descriptions (e.g. “without a diagnosis [of] schizophrenia”), we introduced a second version of - in which the negation of any query component was detected using NegEx, an SVM,, or LingScope, respectively. Further details, rationale, and examples are provided in Supplementary Material Appendix A.
Figure 3.

Overview of the different approaches for (a) query construction and (b) query expansion used for feature extraction in the learning patient cohort retrieval system.

Overview of the different approaches for (a) query construction and (b) query expansion used for feature extraction in the learning patient cohort retrieval system. Query expansion. Figure 3(b) lists the 5 query expansion methods implemented within the L-PCR system. The first 4 query expansion methods incorporate synonyms from UMLS, related concepts from SNOMED CT, synonyms and misspellings from Wikipedia, and individual words from related patient visits using PRF. The fifth query expansion method is the combination of -. Further details and examples are provided in Supplementary Material Appendix B. Note: in addition to query construction and expansion, we extracted any age or gender criteria from the cohort description using a grammar and lexicon previously described in Goodwin et al. and described in Supplementary Material Appendix C.

Electronic health record processing

Stream processing. We unified indexing, searching, and feature extraction across both EHR collections, by representing each EHR as a set of multiple, abstract streams of unstructured information. Each stream corresponds to one or more sections in the EHR collection. Conceptually, each stream acts as a “lens” that determines which sections of the EHR are considered during feature extraction and retrieval. The stream representation allows the L-PCR system to automatically account for the semantics of each stream, without the semantics being explicitly encoded. Figure 4 illustrates the streams used for each EHR collection, while Supplementary Material Appendix D provides additional information about the content of each stream.
Figure 4.

Indexed Streams from EEG reports (left) and hospital records (right). EEG: electroencephalography.

Indexed Streams from EEG reports (left) and hospital records (right). EEG: electroencephalography. Stream indexing. To expedite feature extraction from the EHRs associated with each hospital visit, we separately indexed the content of each EHR collection using Apache Lucene. We used a tiered indexing approach in which each stream was indexed independently, allowing individual streams of each EHR to be retrieved during feature extraction and retrieval. No pre-processing was applied beyond tokenization with Lucene’s English Analyzer.

Visit retrieval

To reduce complexity and improve scalability of the L-PCR system, rather than extracting features from every EHR in the collection, we rely on a basic retrieval step to identify a high-recall set of “candidate” visits likely to be relevant to the cohort description. These candidate visits are obtained by constructing a query with Bag-of-Words (), expanding by all expansions (), and identifying the top ranked EHRs by the All Text stream (/) with the BM25 ranking function (in our experiments we used ). This allowed the set of “candidate” visits to be obtained by mapping the retrieved EHRs to their corresponding patient visits.

Feature extraction

Determining whether a “candidate” patient visit is relevant to (i.e. satisfies the criteria from) a given cohort description requires access to a rich set of features derived from (1) the cohort description , (2) the patient visit , and (3) the interactions between and . To account for the variation between cohort descriptions, we considered multiple strategies for transforming into queries. Let represent the query obtained when using query construction method and query expansion method . Likewise, we considered multiple strategies for representing the information encoded in each visit . Hence, we considered the textual content provided by stream of the electronic health record , and define as the content of stream from each report associated with visit . We produced a single feature vector encoding information about and by extracting the 14 high-level multivalued features listed in Table 2.
Table 2.

Features extracted for a cohort description and hospital visit

Feature descriptionDomain of values
Features encoding information about the cohort descriptiondi
F1Number of criteria detected in cohort description di with each construction method cNC
F2Number of terms in qic,e for each cC, and each expansion method eENC×E
F3Statistics of the normalized inverse document frequency (IDF) of qic,e in each stream sS for each c,e.RA×C×E×S
Features encoding information about the candidate visitvj
F4Number of reports associated with vjN
F5Distribution of report types associated with vjRT
F6Statistics of the number of words in each rsvjs for every sNA×S
Features encoding the relationship between the cohort descriptiondiand candidate visitvj
F7Whether the age (if any) specified in cohort description i matches the age in any stream of any report rsvj0,1
F8Whether the gender (if any) specified in cohort description i matches the most frequently-mentioned gender in any stream of any report rsvj0,1
F9Whether the hospital status in cohort description i matches the hospital status in any stream of any report rsvj0,1
F10Statistics of the Dirichlet-smoothed language model similarity37(LM: Dir) between qic,e and each rsvj for every c,e,sRA×C×E×S
F11Statistics of the Jelinek-Mercer-smoothed language model similarity37(LM: JM) between qic,e and each rsvj for every c,e,sRA×C×E×S
F12Statistics of the BM25 similarity38 between qic,e and each rsvj for every c,e,sRA×C×E×S
F13Statistics of the TF-IDF similarity7 between qic,e and each rsvj for every c,e,sRA×C×E×S
F14Statistics of the term frequency (TF) between qic,e and each rsvj for every c,e,sRA×C×E×S

Additional details for each feature are provided in Supplementary Material Appendix E. represents the natural numbers, represents the real numbers, and the exponent (if provided) indicates the dimensionality, or number of values produced by that feature in the resultant feature vector).

Features extracted for a cohort description and hospital visit Additional details for each feature are provided in Supplementary Material Appendix E. represents the natural numbers, represents the real numbers, and the exponent (if provided) indicates the dimensionality, or number of values produced by that feature in the resultant feature vector). As shown, 10 of the 14 features illustrated in Table 2 are multivalued, i.e. consist of distinct values for each possible query representation of and each stream of (where applicable). Each of these values corresponds to a single entry in the resultant feature vector, i.e. corresponds to 5 entries in the generated feature vector. Moreover, features , and - capture the distribution of feature values extracted for each component of the query () or for each report associated with the hospital visit (, -) using 5 aggregation methods (described below). Of note are features -, which incorporate standard relevance models from IR to measure the relevance between the criteria in and each stream of visit . Aggregation methods. To capture the distribution of feature values obtained using different streams or for each report associated with a candidate visit, we considered 5 aggregating statics .

The learning relevance model

The role of the LRM is to infer a relevance scorebetween every candidate visit and the cohort description using the feature vector extracted above. This is accomplished by using the pairwise strategy of learning-to-rank. Given (1) feature vectors associated with candidate visits and (2) “gold-standard” relevance judgments indicating the relevance of each candidate visit to , the RF is trained to infer the scores which result in the optimal ordering of hospital visits as indicated by . Additional information about the pairwise learning-to-rank strategy, the RF, and model parameters is provided in Supplementary Material Appendix F. After training, the LRM uses the RF to produce the final ranked list of hospital visits by (1) inferring the relevance score for each candidate visit retrieved for and (2) returning the highest scoring visits (in our experiments we used ).

RESULTS

We evaluated the performance of the L-PCR system when automatically identifying patient cohorts in 2 settings: (1) a neurology-specific setting operating exclusively on EEG reports and (2) a general hospital setting associated with a variety of EHR types. In both settings, we measured the performance of our approach using 5 measures commonly used to evaluate IR systems: (1) the Mean Average Precision (MAP), (2) the Normalized Discounted Cumulative Gain (NDCG), (3) the Binary Preference (BPref), (4) the -Precision (rPrec), and (5) the Precision within the first 10 returned visits (P@10); details and discussion of these metrics are provided in Supplementary Material Appendix G.

Patient cohort retrieval from EEG reports

We measured the performance of L-PCR system for neurology-focused cohorts by considering EEG reports from the TUH EEG Corpus using the relevance judgments described in Datasets and Experimental Settings. Table 3(a) presents the performance of the L-PCR system compared to (1) a BM25 baseline and (2) a text-only variant of MERCuRY (a multi-modal retrieval system incorporating polarity information and the BM25 ranking function) using cross validation at the cohort description level (e.g. we evaluated on each set of 3 cohort descriptions when training on the remaining 27). Note: the cohort descriptions used to evaluate MERCuRY in this study are different and more complex than those used in Goodwin and Harabagiu.
Table 3.

Patient cohort retrieval performance on (a) EEG reports and (b) TRECMed

SettingMAPNDCGBPrefrPrecP@10
(a) Retrieval performance when retrieving patient cohorts from EEG reports
 BM25 baseline: 10-fold CV0.49960.61440.40640.52130.6
 L-PCR: 10-fold CV0.66340.71710.59000.60880.6
 MERCuRY (text-only): 10-fold CV0.52200.54410.44830.50810.5
(b) Retrieval performance when retrieving the patient cohorts using in TRECMed from Hospital Records.
 BM25 baseline: evaluated on 20110.40520.52020.50820.41120.600
 BM25 baseline: evaluated on 20120.29300.34620.34620.31350.464
 L-PCR: 10-fold CV on 20110.63160.88160.57880.58590.706
 L-PCR: 10-Fold CV on 20120.51000.81940.47030.50280.589
 L-PCR: trained on 2012 and evaluated on 20110.61270.86750.56380.57630.674
 L-PCR: trained on 2011 and evaluated on 20120.51450.81670.47350.50720.596
 Best submitted to TRECMed 20110.55020.44000.656
 Best submitted to TRECMed 20120.28600.57800.592
Patient cohort retrieval performance on (a) EEG reports and (b) TRECMed

Patient cohort retrieval from hospital EHRs

We evaluated the performance of our approach in a general hospital setting using the patient cohort descriptions produced for TRECMed during 2011 and 2012 (described in Datasets and Experimental Settings). We evaluated the performance of the L-PCR system using the cohort descriptions produced during both years of the evaluation with 10-fold cross validation (using the same strategy described above), as well as when training on the cohort descriptions produced during 2011 and testing on the cohort descriptions produced in 2012 (and vice versa). Table 3(b) presents these results, as well as performance of (1) a BM25 baseline system and (2) the best submitted systems reported by NIST for each year, which to the best of our knowledge are still state-of-the-art. An overview of these systems is provided in Supplementary Material Appendix J.

Measuring feature importance

To analyze the impact of the different techniques and features used by the L-PCR system, we measured the Gini importance of each entry in all feature vectors extracted from the training data. Figure 5 illustrates the normalized average Gini importance of different (1) query construction methods, (2) query expansion methods, (3) features, (4) aggregating statistics, or (5/6) streams for each collection of EHRs. Supplementary Material Appendix I provides the most important features for each experiment.
Figure 5.

Average feature importance as measured using EEG reports and TRECMed hospital records. EEG: electroencephalography.

Average feature importance as measured using EEG reports and TRECMed hospital records. EEG: electroencephalography.

DISCUSSION

In this article, we presented a learning PCR system which incorporates machine learning operating on features encoding thousands of different strategies for representing queries, EHRs, and their interactions to learn how to rank patient cohorts based on clinicians’ feedback. As shown by Table 3, the L-PCR surpassed previously published state-of-the-art NDCG score by 27.1% on EEGs and by 52.5% on TRECMed 2011. It is interesting to note that for both years of TRECMed, there was no statistically significant change in performance when trained using cross validation, or using the cohort descriptions from the previous/following year (). It is clear from Figure 5(a) that bag-of-words obtained the best performance compared to other query construction methods when processing cohorts from EEGs. By contrast, for general EHRs, Wikipedia titles provided higher performance. This reinforces our observations that many neurological phenomena (e.g. “spike and wave”) are not associated with entities in structured knowledge bases. Moreover, the fact that, for EEGs, Wikipedia redirects of individual words in the cohort description provided the most informative expansions suggests that when structured information is available, it greatly improves performance. This is clearly demonstrated by the TRECMed cohorts, where the most impactful query expansion method was based on UMLS, followed closely by PRF. Interestingly, the impact of each individual features, as shown in Figure 5(c), varied between all 3 experiments, suggesting that the choice of relevance model is less important than the choice of query composition and expansion strategies. Overall, we found that query expansion resulted in a 6.0% (relative) increase in performance compared to no query expansion for TRECMed, and an 8.7% increase for EEG reports. One key difference between automatic PCR systems and traditional IR systems is that each patient may be associated with multiple medical records. When analyzing different methods of aggregating features from report- to visit-level, the importance of the variance statistic, especially when contrasted with the sum statistic, suggests that the reports associated with a visit often have unequal impacts on the relevance of the visit. Moreover, as the sum measure closely resembles the effect of merging all reports associated with a visit into a single document (a common strategy employed by TRECMed participants), our results show that treating reports separately can be advantageous. In terms of streams, as shown in Figure 5(e–f), the AllText stream provided the most information to the model, suggesting that differentiating between streams added little value to the L-PCR system.

Error analysis

We analyzed errors made by the L-PCR system for both EHRs collection. In the TUH EEG collection, we observed several common phenomena. The first was that many of the cohort criteria were not found in existing ontologies and were often incorrectly parsed. For example, “PLED” indicates “periodic lateralized epileptiform discharges”, but was not present in UMLS, SNOMED, Wikipedia, or even the Epilepsy and Seizure Ontology (EpSO). A more significant source of errors was accounting for the fact that EEG cohorts are typically characterized by the attributes of waveform activity, rather than their presence or absence. In the previous example the “lateral” attribute highlights a major phenomenon in EEG reports—the role of spatiotemporal information. In EEG reports, the term “lateral” may not be mentioned; instead, activity is often described as occurring in specific lobes of the brain or at specific channels/nodes in the EEG. Moreover, consider that for a visit to be relevant to the criteria “generalized periodic epileptiform discharges”, it is not sufficient for “epileptiform discharges” to be described. The discharges must be generalized and periodic. Each of these attributes can be described in multiple ways, for example, both “non-localized”, “diffuse” indicate generalized activity. We also observed in both collections that the performance of the L-PCR (and each baseline system) varied significantly between cohort descriptions. We found that cohort descriptions which qualified their criteria with anatomical (e.g. “lower extremity”) or temporal (“history of”) attributes were harder to retrieve. Moreover, we observed that the most difficult cohort descriptions were those that described relations between concepts: i.e. “inflammatory disorders receiving TNF-inhibitor treatment” or “cancer patients with liver metastasis treated in the hospital who underwent a procedure”. For an in-depth analysis of errors encountered by participants of the TRECMed evaluation, we refer the reader to Edinger et al.

Limitations

This study has several limitations. First, only 111 cohort descriptions were evaluated. While every effort was made to produce and evaluate realistic patient cohort descriptions, the time and cost associated with producing relevance judgments limited the number of cohorts that could be evaluated. Consequently, the performance results reported in Table 3 represent results of a pilot study only on 2 sets of EHR collections and may not be reflective of performance on other EHR collections. Second, many of the features extracted in our study rely on individual streams in the EEG/hospital reports which vary between hospitals and EHR collections. Fortunately, in our experiments, the top-performing features all relied on the AllText stream, which can be easily generalized across EHR collections. Finally, because the TRECMed collection is no longer available, we were unable to consider some promising features, such as word embeddings or convolutional neural network features. In future work, we shall investigate the impact of concept embeddings on the performance of the L-PCR system.

CONCLUSION

Learning-to-rank can be successfully applied for retrieving patient cohorts from EHR when (1) judgments of relevance are available; and (2) a rich set of features is considered. In this paper, we present L-PCR system, and our experimental results on 2 EHR collections demonstrate that the L-PCR system obtains results 27–53% above state-of-the-art PCR systems when retrieving the same cohorts from the same EHR collections. Moreover, by analyzing the performance of the L-PCR system, we were able to measure the impact of a variety of PCR techniques. Overall, our results indicate the promise of the L-PCR system, but also reveal potent avenues for further improvement. Conflict of interest statement. The authors have no conflicts of interest to declare.

CONTRIBUTORS

T.G. and S.H. originated the study. T.G. conducted the experiments. S.H. reviewed and help analyze the findings. T.G. and S.H. wrote the first and revised drafts of the manuscript.

FUNDING

Research reported in this publication was supported by the National Human Genome Research Institute of the National Institutes of Health under award number 1U01HG008468. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. Click here for additional data file.
  15 in total

1.  SNOMED clinical terms: overview of the development process and project status.

Authors:  M Q Stearns; C Price; K A Spackman; A Y Wang
Journal:  Proc AMIA Symp       Date:  2001

2.  Multi-modal Patient Cohort Identification from EEG Report and Signal Data.

Authors:  Travis R Goodwin; Sanda M Harabagiu
Journal:  AMIA Annu Symp Proc       Date:  2017-02-10

Review 3.  Unified EEG terminology and criteria for nonconvulsive status epilepticus.

Authors:  Sándor Beniczky; Lawrence J Hirsch; Peter W Kaplan; Ronit Pressler; Gerhard Bauer; Harald Aurlien; Jan C Brøgger; Eugen Trinka
Journal:  Epilepsia       Date:  2013-09       Impact factor: 5.864

4.  A flexible framework for deriving assertions from electronic medical records.

Authors:  Kirk Roberts; Sanda M Harabagiu
Journal:  J Am Med Inform Assoc       Date:  2011-07-01       Impact factor: 4.497

5.  Improving search over Electronic Health Records using UMLS-based query expansion through random walks.

Authors:  David Martinez; Arantxa Otegi; Aitor Soroa; Eneko Agirre
Journal:  J Biomed Inform       Date:  2014-04-21       Impact factor: 6.317

6.  MetaMap is a superior baseline to a standard document retrieval engine for the task of finding patient cohorts in clinical free text.

Authors:  K Bretonnel Cohen; Tom Christiansen; Lawrence E Hunter
Journal:  Text Retr Conf       Date:  2011

7.  The Unified Medical Language System.

Authors:  D A Lindberg; B L Humphreys; A T McCray
Journal:  Methods Inf Med       Date:  1993-08       Impact factor: 2.176

8.  Automated disease cohort selection using word embeddings from Electronic Health Records.

Authors:  Benjamin S Glicksberg; Riccardo Miotto; Kipp W Johnson; Khader Shameer; Li Li; Rong Chen; Joel T Dudley
Journal:  Pac Symp Biocomput       Date:  2018

9.  Epilepsy and seizure ontology: towards an epilepsy informatics infrastructure for clinical research and patient care.

Authors:  Satya S Sahoo; Samden D Lhatoo; Deepak K Gupta; Licong Cui; Meng Zhao; Catherine Jayapandian; Alireza Bozorgi; Guo-Qiang Zhang
Journal:  J Am Med Inform Assoc       Date:  2013-05-18       Impact factor: 4.497

Review 10.  A review of approaches to identifying patient phenotype cohorts using electronic health records.

Authors:  Chaitanya Shivade; Preethi Raghavan; Eric Fosler-Lussier; Peter J Embi; Noemie Elhadad; Stephen B Johnson; Albert M Lai
Journal:  J Am Med Inform Assoc       Date:  2013-11-07       Impact factor: 4.497

View more
  4 in total

1.  Patient Cohort Retrieval using Transformer Language Models.

Authors:  Sarvesh Soni; Kirk Roberts
Journal:  AMIA Annu Symp Proc       Date:  2021-01-25

2.  Medical knowledge infused convolutional neural networks for cohort selection in clinical trials.

Authors:  Chi-Jen Chen; Neha Warikoo; Yung-Chun Chang; Jin-Hua Chen; Wen-Lian Hsu
Journal:  J Am Med Inform Assoc       Date:  2019-11-01       Impact factor: 4.497

3.  Evaluation of patient-level retrieval from electronic health record data for a cohort discovery task.

Authors:  Steven R Chamberlin; Steven D Bedrick; Aaron M Cohen; Yanshan Wang; Andrew Wen; Sijia Liu; Hongfang Liu; William R Hersh
Journal:  JAMIA Open       Date:  2020-07-26

4.  Test collections for electronic health record-based clinical information retrieval.

Authors:  Yanshan Wang; Andrew Wen; Sijia Liu; William Hersh; Steven Bedrick; Hongfang Liu
Journal:  JAMIA Open       Date:  2019-06-04
  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.