Literature DB >> 28651340

Linking quality indicators to clinical trials: an automated approach.

Enrico Coiera1, Miew Keen Choong1, Guy Tsafnat1, Peter Hibbert1,2, William B Runciman1,2,3.   

Abstract

OBJECTIVE: Quality improvement of health care requires robust measurable indicators to track performance. However identifying which indicators are supported by strong clinical evidence, typically from clinical trials, is often laborious. This study tests a novel method for automatically linking indicators to clinical trial registrations.
DESIGN: A set of 522 quality of care indicators for 22 common conditions drawn from the CareTrack study were automatically mapped to outcome measures reported in 13 971 trials from ClinicalTrials.gov. INTERVENTION: Text mining methods extracted phrases mentioning indicators and outcome phrases, and these were compared using the Levenshtein edit distance ratio to measure similarity. MAIN OUTCOME MEASURE: Number of care indicators that mapped to outcome measures in clinical trials.
RESULTS: While only 13% of the 522 CareTrack indicators were thought to have Level I or II evidence behind them, 353 (68%) could be directly linked to randomized controlled trials. Within these 522, 50 of 70 (71%) Level I and II evidence-based indicators, and 268 of 370 (72%) Level V (consensus-based) indicators could be linked to evidence. Of the indicators known to have evidence behind them, only 5.7% (4 of 70) were mentioned in the trial reports but were missed by our method.
CONCLUSIONS: We automatically linked indicators to clinical trial registrations with high precision. Whilst the majority of quality indicators studied could be directly linked to research evidence, a small portion could not and these require closer scrutiny. It is feasible to support the process of indicator development using automated methods to identify research evidence.
© The Author 2017. Published by Oxford University Press in association with the International Society for Quality in Health Care.

Entities:  

Keywords:  clinical trials; concept mapping; indicator; quality of health care; text mining

Mesh:

Year:  2017        PMID: 28651340      PMCID: PMC5890874          DOI: 10.1093/intqhc/mzx076

Source DB:  PubMed          Journal:  Int J Qual Health Care        ISSN: 1353-4505            Impact factor:   2.038


Introduction

Clinical indicators are important tools for assessing the quality of health care, and for identifying and prioritizing areas for improvement [1, 2]. To be effective, such indicators need to be robust measures of system performance that correlate with the processes of interest, and be cost-effective to measure. A lack of uniformity in reporting the rationale for selecting indicators means that it can prove difficult to know whether a given indicator is based on research evidence. Deductive development of indicators is the most common approach taken [3], where clinical indicators are extracted from clinical guidelines or are identified in the process of guideline development [4]. Most clinical guidelines are based on systematic reviews which are syntheses, mainly of randomized controlled trials (RCTs) [5]. Thus, the development of indicators typically requires a lengthy manual process of searching for and analysis of the research evidence underpinning guidelines. The development of methods that assist in identifying candidate indicators from the research evidence, or that validate existing indicators against the evidence base, should help increase the efficiency of indicator development, and may also improve confidence in the quality of studies based on indicators. The development of robust and reliable ways of linking indicators to the evidence base is key to achieving this. In a previous study, focussing on paediatric asthma, we demonstrated that manual linking of indicators to clinical outcomes in trial reports found a link to research for 95% of standard indicators [6]. The emergence of new methods such as computational text mining is allowing other complex processes, such as the creation of systematic reviews of the research literature, to be automated [7]. By breaking down the steps in such a complex manual process, it is possible to identify individual steps in which computational tools can assist or replace manual work, improving either efficiency or accuracy [8]. The process of indicator development can similarly be reduced to several steps that form a developmental pipeline [3, 4]. This indicator development pipeline begins with the selection of a clinical or health service process that needs to be monitored, and continues with the identification of candidate indicators, their appraisal based upon performance criteria, and then implementation and evaluation (Fig. 1).
Figure 1

The indicator development pipeline. For a disease or health service process that requires monitoring, appropriate measurements or indicators are required. Such indicators can be identified by statistical analyses of electronic health records, reviews of outcome measures used in clinical trials or clinical practice guidelines, or in the absence of strong evidence, from expert recommendations. The selection of indicators from amongst these candidates is aided by evidence of the indicator's predictive performance as a measure of the process in question—taken from research or record analysis, along with technical and economic evidence about the feasibility of using the indicator in practice, and any necessary expert views. Once implemented, additional data can be gathered to update assessments of an indicator's performance and real-world feasibility.

The indicator development pipeline. For a disease or health service process that requires monitoring, appropriate measurements or indicators are required. Such indicators can be identified by statistical analyses of electronic health records, reviews of outcome measures used in clinical trials or clinical practice guidelines, or in the absence of strong evidence, from expert recommendations. The selection of indicators from amongst these candidates is aided by evidence of the indicator's predictive performance as a measure of the process in question—taken from research or record analysis, along with technical and economic evidence about the feasibility of using the indicator in practice, and any necessary expert views. Once implemented, additional data can be gathered to update assessments of an indicator's performance and real-world feasibility. The aim of this study is to assess the degree to which is possible to use computational text processing methods to assist in one step in the indicator development pipeline—to automatically link candidate indicators to published clinical trial registrations, to assist with indicator appraisal and selection. Our approach is to seek links between candidate indicators and the outcome measures reported for clinical trials. The rationale for this approach is that the evidence for the effectiveness of an indicator is likely captured in clinical trials that use the indicator as an outcome measure. To test the generalizability of the approach we looked to the CareTrack study [9], which measured the quality of care provided for 22 common health conditions, using 522 different indicators.

Methods

Linking indicators to outcome measures in the published evidence requires a mapping to be developed. We developed a simple text-processing pipeline that takes a given indicator and attempts to map it to a collection of clinical trial reports. To evaluate the accuracy of this computational method we undertook an evaluation study that consisted of four steps: Creation of a test set of candidate indicators. Creation of a list of candidate outcome measures from a test set of RCTs. Automatically mapping indicator and outcome measures in these two sets. An analysis of the outcome of mapping clinical indicators to outcome measures using the text processing method.

Creation of a list of candidate indicator measures

To create a candidate list of indicator measures, we looked to the CareTrack Australia study [9]. An expert-driven process identified 522 clinical quality indicators across the 22 conditions. Various levels of evidence were provided for each CareTrack indicator. The majority of indicators (71%) were Level V (consensus-based). The rest were either associated with Levels I and II (Level I—Systematic Reviews and Level II—Randomized Controlled Trials) evidence (13%) or Levels III and IV evidence [5] (16%). We automatically extracted the CareTrack indicators from the list published in Appendix 1 of the main study [9]. This produced a list of noun phrases (called ‘indicator phrases’ hereafter). At least one indicator phrase was found for each of the 522 indicators. The method used purpose built rules that located sections of sentences in a text that were likely to contain mention of an indicator or outcome measure. Next the method filtered obviously incorrect terms, and normalized the remaining indicator phrases to a standard common format (see Appendix 1, Section 1).

Creation of a list of candidate outcome measures from RCT

To create a list of candidate outcome measures from research trials, we used ClinicalTrials.gov, a web-based clinical trial registry. The United States Food and Drug Administration Amendments Act (FDAAA 801) requires clinical studies of FDA regulated products to be prospectively registered in CliniclTrials.gov. The International Committee of Medical Journal Editors also requires prospective registration of clinical trials as a prerequisite for publication [10]. Registration includes recording the clinical trial, its methods and measured outcomes before trial commencement and reporting the results after the trial is concluded [11]. ClinicalTrials.gov specifically provides links between its registry entries and published results in research articles using a unique identifier (the NCT Number) for each study [12, 13]. To date, there are over 34 000 trials registered on ClinicalTrials.gov. Whilst ClinicalTrials.gov is not a complete list of all trials or their published results, it is large enough to act as a test resource for the present study. We used the advanced search feature of ClinicalTrials.gov to search for the 22 conditions studied in CareTrack with a publication date before 31 December 2010, to mirror the period over which the indicators were developed. Only trials that used randomized allocations, parallel/crossover/factorial intervention design (excluding single group intervention design), and that reported outcome measures, were included. A total of 13 971 trials met these inclusion criteria. We automatically extracted the noun phrases from the outcome measures recorded for the included trials with the same text processing method used to extract indicators, to create a list of ‘outcome phrases’ (see Appendix 1, Section 2).

Automatically mapping indicator and outcome measures between the two sets

Indicator phrases and outcome phrases were then pooled, and placed into clusters if the Levenshtein edit distance ratio between phrases was 75% or greater [14]. All the phrases in a given cluster were considered mapped to each other.

Analysis of mapping outcomes

We used both a strict and a lenient evaluation method. Under lenient evaluation, an indicator phrase was labelled ‘mapped’ if any of the corresponding indicator phrases could be mapped to any outcome phrases or ‘miss’ otherwise. Under strict evaluation, an indicator was only labelled ‘mapped’ if all of its corresponding indicator phrases could be linked to outcome phrases from ClinicalTrials.gov. To establish a benchmark for the effectiveness of the phrase extraction and mapping pipeline, it was validated against a human gold standard. One hundred randomly selected clinical indicators and 100 randomly selected RCT outcome measures were manually mapped to create the gold standard. Precision, recall and F1 scores were calculated for the performance of the automated method against this gold standard (Table 1). The precision of the mapping between indicator and outcome phrases was high at 0.88.
Table 1

Performance of the lexical pipeline for extraction of indicator and outcome phrases and for mapping each to the other, against a gold standard validation set

RecallPrecisionF1 score
Indicator phrases0.930.510.66
Outcome phrases0.980.640.77
Mapping0.850.880.86
Performance of the lexical pipeline for extraction of indicator and outcome phrases and for mapping each to the other, against a gold standard validation set

Results

Indicator and outcome phrase extraction

The automated method was able to extract indicator and outcome phrases for every indicator and trial (Tables 2 and 3). For the 522 indicators, an average of 2.8 phrases were extracted per indicator (1.8 unique phrases). For the 13 971 trials from ClinicalTrials.gov, an average of 2.5 outcome phrases were extracted (0.71 unique phrases) per outcome measure.
Table 2

Extraction and mapping results for clinical indicators and clinical trial outcome measures in 22 health conditions

ConditionNumber of indicatorsNumber of extracted indicator phrasesNumber of trialsNumber of extracted outcome phrasesNumber (%) indicators linked to outcome phrases (lenient)Number (%) indicators linked to outcome phrases (strict)
Alcohol dependence1338248224112 (92.3)7 (53.8)
Antibiotic use5511945 (100)5 (100)
Asthma286594713 59125 (89.3)15 (53.6)
Atrial fibrillation1843333428712 (66.7)4 (22.2)
Cerebrovascular accident3597588690828 (80.0)12 (34.3)
Chronic heart failure42109188227137 (88.1)8 (19.1)
Chronic obstructive pulmonary disease3910667711 11229 (74.4)15 (38.5)
Community acquired pneumonia33110517957 (21.2)2 (6.1)
Coronary artery disease38127145819 32536 (94.7)20 (52.6)
Depression1954130314 67514 (73.7)9 (47.4)
Diabetes2768321542 09225 (92.6)13 (48.1)
Dyspepsia2264636055 (22.7)1 (4.5)
Hyperlipidaemia1535491641610 (66.7)3 (20.0)
Hypertension4998154616 94442 (85.7)26 (53.1)
Low back pain104123324825 (50.0)0 (0)
Obesity924132613 3018 (88.9)4 (44.4)
Osteoarthritis2170631772017 (81.0)6 (28.6)
Osteoporosis102139762147 (70.0)2 (20.0)
Panic disorder1465516754 (28.6)0 (0)
Preventive care3169221956 (19.4)0 (0)
Surgical site infection57676375 (100)4 (80.0)
Venous thromboembolism39139125200214 (35.9)1 (2.6)
Total522145513 971174 582353 (67.6)157 (30.1)
Table 3

Examples extracted phrases

MeasureIndicator phrases (CareTrack)Outcome phrases (ClinicalTrials.gov)
Forced expiratory volume in 1 s (FEV1)Expiratory volume in 1 s (FEV1)

forced expiratory volume in one second (FEV1);

Forced Expiratory Volume in 1 second (FEV1);

expiratory volume in 1 second (FEV1);

FEV1 (Forced Expiratory Volume in 1 Second);

Forced Expiratory Volume in 1 s(FEV-1);

Forced Expiratory Volume in the first second (FEV1);

Forced expiratory volume in one-second (FEV1);

Volume in 1 sec (FEV1);

expiratory volume (FEV1);

Spirometry Forced Expiratory Volume in One Second (FEV1);

Forced Expiratory Volume in 1 second (FEV1): spirometry

Glycosylated hemoglobin (HbA1c)Glycated hemoglobin (HBA1c) levels

Glycosylated hemoglobin (HbA1c) levels;

glycosolated haemoglobin;

Glycated hemoglobin (HbA1c);

HbA1c (Glycosylated hemoglobin);

HbA1C: Glycated Hemoglobin;

Hemoglobin (HbA1c);

Glcosylated Hemoglobin (HbA1c);

HbA1c Test (Glycated hemoglobin);

Glycosylated Hemoglobin A1c (HbA1c)

Extraction and mapping results for clinical indicators and clinical trial outcome measures in 22 health conditions Examples extracted phrases forced expiratory volume in one second (FEV1); Forced Expiratory Volume in 1 second (FEV1); expiratory volume in 1 second (FEV1); FEV1 (Forced Expiratory Volume in 1 Second); Forced Expiratory Volume in 1 s(FEV-1); Forced Expiratory Volume in the first second (FEV1); Forced expiratory volume in one-second (FEV1); Volume in 1 sec (FEV1); expiratory volume (FEV1); Spirometry Forced Expiratory Volume in One Second (FEV1); Forced Expiratory Volume in 1 second (FEV1): spirometry Glycosylated hemoglobin (HbA1c) levels; glycosolated haemoglobin; Glycated hemoglobin (HbA1c); HbA1c (Glycosylated hemoglobin); HbA1C: Glycated Hemoglobin; Hemoglobin (HbA1c); Glcosylated Hemoglobin (HbA1c); HbA1c Test (Glycated hemoglobin); Glycosylated Hemoglobin A1c (HbA1c)

Indicator to outcome mapping

An average of 23 outcome phrases (IQR = 9) were linked to each indicator phrase. Using the strict evaluation criterion, it was possible to link all phrases associated with an indicator to one or more outcome phrases for 157/522 (30%) of indicators. Using lenient evaluation, relaxing the mapping criterion to require only one or more mappings per indicator phrase, added an additional 196 (38%) indicators, bringing the total number of indicators with a mapping to a clinical trial to 68%. The remaining 169 (32%) indicators could not be mapped to any outcome phrase. There were 70 CareTrack indicators known to be associated with Level I or II evidence and 71% (50/70) of these were mapped to one or more outcome phrases in clinical trials (or 37% using the strict criterion of full mapping). A further 370 indicators were understood to be consensus-based in the original CareTrack study. Amongst these, 72% (268 of 370) could be mapped to a trial (and 33% were strictly mapped). Figure 2 shows the success in mapping CareTrack indicators by level of evidence. On average, 21 studies were mapped to an indicator with Levels I and II evidence and 24 studies linked to a consensus-based indicator (Fig. 3).
Figure 2

Percentage of indicator phrases mapped to clinical trials, by level of available evidence as understood by the indicator developers.

Figure 3

Average number of clinical trials found per indicator, by level of available evidence as assessed by the indicator developers.

Percentage of indicator phrases mapped to clinical trials, by level of available evidence as understood by the indicator developers. Average number of clinical trials found per indicator, by level of available evidence as assessed by the indicator developers.

Error analysis

There were 20/70 (29%) Levels I and II evidence-based indicators that could not be mapped to any outcome phrase. Manual analysis of mapping failures revealed that seven of the failed indicators appeared to be true negatives, in that manual methods were also unable to identify any clinical trial in the trial test set which contained matching outcome measures. Four indicators (5.7%) were false negatives, i.e. mappings to outcome phrases were possible but were missed by our method. For the final nine indicators, outcome phrases existed but were located in the intervention field of the clinical trial record in ClinicalTrials.gov, instead of the outcome measure field.

Discussion

Indicator development is an important, but currently difficult and manual process. The indicator development process appears well suited to automation in many, potentially all, of its stages. In this paper we have focussed on exploring the feasibility of automating just one stage in the overall indicator development pipeline—the identification of candidate indicators and the research that may support their use. Indeed, this appears to be the first study we are aware of to automatically link clinical indicators with clinical trial registrations. Only 13% of the original CareTrack indicators were identified by their source guidelines to have Level I or II evidence behind them and 16% had Level III or IV evidence. The remaining 71% were labelled as being supported by expert consensus. Our methods nevertheless identified candidate clinical trials (Level II evidence) for 72% of the consensus indicators, 67% of the Level IV and 63% of Level III indicators. The CareTrack process relied on clinical guidelines to identify the level of evidence associated with a given indicator. This suggests that, despite best practice manual methods, important research evidence may be missed, either when developing guidelines, or relying on guidelines. Whilst this study did not further examine the quality of those studies, they do meet the criteria for Level I or II evidence in their clincialtrial.gov record. This is promising, as it suggests that automated methods will be able to provide a more rigorous and much more efficient approach to mapping indicators to evidence. Our approach used simple, standard text processing methods to extract and normalize candidate phrases in documents containing mentions of indicators or clinical trial measures. Despite the simplicity of our approach in this feasibility study, initial validation of the methods against a gold standard showed the method performed with a precision of 0.88 and F1 score of 0.86. When applied to the 70 CareTrack indicators known to have Level I or II evidence, some link to clinical trial registrations was possible for 73% of Level I and 67% of Level II indicators. To understand these results, it should be noted that there are two major reasons for failure to find a mapping. The first is that no mapping exists because there is no available trial registration to support the indicator. The second is that a mapping does exist but the mapping method fails to find it. Detailed analysis of the performance for Levels I and II indicators showed that for 10% no mapping was possible within the given set of trials, and for another 6% the method failed. In another 13% mapping was possible but the trial report itself was the problem, incorrectly reporting outcome measures in the wrong field. Interestingly, the indicators that had poor mapping appeared to have a lower number of trials in the ClinicalTrials.gov test set compared to other indicators, suggesting that sample size may have contributed to poorer performance using the current methods. Figure 4 shows that there is an association between mapping success and the number of trials available to be mapped.
Figure 4

Scatter plot showing the association between success in finding indicators for the 22 CareTrack conditions and the sample size of RCTs available for that condition in the test corpus. A logarithmic regression function is fitted to the data.

Scatter plot showing the association between success in finding indicators for the 22 CareTrack conditions and the sample size of RCTs available for that condition in the test corpus. A logarithmic regression function is fitted to the data. Performance is likely to improve with more robust methods. For example, rather than mapping raw phrases using an edit distance, the phrases could be semantically labelled using standard tools such as MMTX which comes with the UMLS metathesaurus [15]. For the overall indicator development pipeline to be supported, additional work is required at each of the stages identified in Fig. 1. For example, text-mining tools can be used to extract indicators not just from trial registrations, but also from randomized clinical trial reports, systematic reviews and clinical practice guidelines. We also did not examine the results of the clinical trials that mapped to indicators. We did not undertake an assessment of whether the mapped trials provide evidence for the use of an indicator in a health settings or whether these indicators are economical and effective to apply. Such considerations are important further stages in the indicator development process, and different methods would be needed to support them. Not all published research associated with a trial is directly linked to the ClinicalTrials.gov registration, and in one study 44% of registrations with no linked publications were found to have published articles after a manual search [16]. While a lack of linkage between trial registrations and the literature did not seem to impede the identification of outcome measures that could serve as indicators, it will probably be necessary to search for these reports when assessing indicator suitability. Finally, little work has been done to utilize the data stored in electronic health records, which can also be used to identify candidate indicators based on their ability to predict specific clinical conditions or events.

Limitations

This study only examined RCTs from one source (ClinicalTrials.gov). A broader range of clinical trial repositories exists and their use would likely identify additional trials relevant to indicator development. Equally we did not search for reports associated with Levels III and IV evidence, which may also be of value during the indicator selection process.

Conclusion

We have presented a method to automatically identify clinical trial reports that may be relevant to the selection of clinical indicators. Whilst the methods used are simple, they appear to identify trials missed by the developers of clinical indicators, and so should prove to be beneficial in the indicator development process.
  15 in total

1.  Effective mapping of biomedical text to the UMLS Metathesaurus: the MetaMap program.

Authors:  A R Aronson
Journal:  Proc AMIA Symp       Date:  2001

2.  Defining and classifying clinical indicators for quality improvement.

Authors:  Jan Mainz
Journal:  Int J Qual Health Care       Date:  2003-12       Impact factor: 2.038

3.  Clinical trial registration: a statement from the International Committee of Medical Journal Editors.

Authors:  Catherine De Angelis; Jeffrey M Drazen; Frank A Frizelle; Charlotte Haug; John Hoey; Richard Horton; Sheldon Kotzin; Christine Laine; Ana Marusic; A John P M Overbeke; Torben V Schroeder; Hal C Sox; Martin B Van Der Weyden
Journal:  N Engl J Med       Date:  2004-09-08       Impact factor: 91.245

4.  Medicine. Moving toward transparency of clinical trials.

Authors:  Deborah A Zarin; Tony Tse
Journal:  Science       Date:  2008-03-07       Impact factor: 47.728

5.  The automation of systematic reviews.

Authors:  Guy Tsafnat; Adam Dunn; Paul Glasziou; Enrico Coiera
Journal:  BMJ       Date:  2013-01-10

6.  CareTrack: assessing the appropriateness of health care delivery in Australia.

Authors:  William B Runciman; Tamara D Hunt; Natalie A Hannaford; Peter D Hibbert; Johanna I Westbrook; Enrico W Coiera; Richard O Day; Diane M Hindmarsh; Elizabeth A McGlynn; Jeffrey Braithwaite
Journal:  Med J Aust       Date:  2012-07-16       Impact factor: 7.738

7.  Publication of NIH funded trials registered in ClinicalTrials.gov: cross sectional analysis.

Authors:  Joseph S Ross; Tony Tse; Deborah A Zarin; Hui Xu; Lei Zhou; Harlan M Krumholz
Journal:  BMJ       Date:  2012-01-03

8.  Comparing clinical quality indicators for asthma management in children with outcome measures used in randomised controlled trials: a protocol.

Authors:  Miew Keen Choong; Guy Tsafnat; Peter Hibbert; William B Runciman; Enrico Coiera
Journal:  BMJ Open       Date:  2015-09-08       Impact factor: 2.692

9.  Linking clinical quality indicators to research evidence - a case study in asthma management for children.

Authors:  Miew Keen Choong; Guy Tsafnat; Peter Hibbert; William B Runciman; Enrico Coiera
Journal:  BMC Health Serv Res       Date:  2017-07-21       Impact factor: 2.655

10.  Systematic review automation technologies.

Authors:  Guy Tsafnat; Paul Glasziou; Miew Keen Choong; Adam Dunn; Filippo Galgani; Enrico Coiera
Journal:  Syst Rev       Date:  2014-07-09
View more
  3 in total

1.  Does health informatics have a replication crisis?

Authors:  Enrico Coiera; Elske Ammenwerth; Andrew Georgiou; Farah Magrabi
Journal:  J Am Med Inform Assoc       Date:  2018-08-01       Impact factor: 4.497

2.  Clinical indicators for common paediatric conditions: Processes, provenance and products of the CareTrack Kids study.

Authors:  Louise K Wiles; Tamara D Hooper; Peter D Hibbert; Charlotte Molloy; Les White; Adam Jaffe; Christopher T Cowell; Mark F Harris; William B Runciman; Annette Schmiede; Chris Dalton; Andrew R Hallahan; Sarah Dalton; Helena Williams; Gavin Wheaton; Elisabeth Murphy; Jeffrey Braithwaite
Journal:  PLoS One       Date:  2019-01-09       Impact factor: 3.240

3.  Clinical Trial Assessment Principles of National Class III Medical Devices in China.

Authors:  Yue Min; Jin-Tao Gao; Jing Wu; Bao Zhai; Dan Han; Bin Liu
Journal:  Orthop Surg       Date:  2019-09-06       Impact factor: 2.071

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.