Literature DB >> 30337312

Cross-sectional analysis of UK research studies in 2015: results from a scoping project with the UK Health Research Authority.

Tim Clark1, Richard H Wicentowski2, Matthew R Sydes3.   

Abstract

OBJECTIVES: To determine whether data on research studies held by the UK Health Research Authority (HRA) could be summarised automatically with minimal manual intervention. There are numerous initiatives to reduce research waste by improving the design, conduct, analysis and reporting of clinical studies. However, quantitative data on the characteristics of clinical studies and the impact of the various initiatives are limited.
DESIGN: Feasibility study, using 1 year of data.
SETTING: We worked with the HRA on a pilot study using research applications submitted for UK-wide ethical review. We extracted into a single dataset, information held in anonymised XML files by the Integrated Research Application System (IRAS) and the HRA Assessment Review Portal (HARP). Research applications from 2014 to 2016 were provided. We used standard text extraction methods to assess information held in free-text fields. We use simple, descriptive methods to summarise the research activities that we extracted. PARTICIPANTS: Not applicable-records-based study
INTERVENTIONS: Not applicable. PRIMARY AND SECONDARY OUTCOME MEASURES: Feasibility of extraction and processing.
RESULTS: We successfully imported 1775 non-duplicate research applications from the XML files into a single database. Of these, 963 were randomised controlled trials and 812 were other studies. Most studies received a favourable opinion. There was limited patient and public involvement in the studies. Most, but not all, studies were planned for publication of results. Novel study designs (eg, adaptive and Bayesian designs) were infrequently reported.
CONCLUSIONS: We have demonstrated that the data submitted from IRAS to the HRA and its HARP system are accessible and can be queried for information. We strongly encourage the development of fully resourced collaborative projects to further this work. This would aid understanding of how study characteristics change over time and across therapeutic areas, as well as the progress of initiatives to improve the quality and relevance of research studies. © Author(s) (or their employer(s)) 2018. Re-use permitted under CC BY. Published by BMJ.

Entities:  

Keywords:  clinical trials

Mesh:

Year:  2018        PMID: 30337312      PMCID: PMC6196875          DOI: 10.1136/bmjopen-2018-022340

Source DB:  PubMed          Journal:  BMJ Open        ISSN: 2044-6055            Impact factor:   2.692


First study to draw information from Health Research Authority (HRA’s) Assessment Review Portal HARP system into one searchable dataset. Anonymised data from the HRA’s HARP system can be interrogated with minimal manual intervention. Feasibility study, so limited to data from 2015 only and excluding phase I healthy volunteer studies. Aligning the questions from the XML document was not straightforward. The search terms chosen for the free-text fields were not exhaustive and because it was unfeasible in a pilot to review all applications, the sensitivity and specificity of the text-mining methodology could not be calculated.

Introduction

The need to improve the quality of clinical research is increasingly understood.1 With a particular focus on clinical studies, the ‘gold standard’ for evidence-based medicine, there has been a noticeable push towards: improving study protocols1; developing and implementing newer methodologies, such as adaptive designs2; involving patients in the design, conduct and management of studies3; and ensuring that study results are quickly and accurately reported.4 In order to properly evaluate the current state of clinical research and changes over time, it is clearly necessary to have unfettered access to the research protocols and study results.5 6 There are a number of limitations in accessing such information. Published evaluations of the state of clinical research are mainly based on the clinical study publication due to the difficulty in obtaining access to unpublished research protocols.7 Research registers are incomplete for many jurisdictions and/or are, like study publications, limited in detail compared with the research protocol. For example, clinicaltrials.gov8 has good coverage of clinical studies in North America, where registration of clinical trials is mandatory. The WHO International Clinical Trials Registry Portal9 provides access to 15 different regional or national registries, but both have markedly less detail than in the study’s research protocol. Published evaluations of the literature are likely affected by publication bias and there are often discrepancies between research publications and their underpinning protocols.10 The ability to analyse a large database of application forms, each of which contain very detailed information taken directly from the research protocol, would enable researchers to perform a detailed examination of the characteristics of clinical studies, particularly those necessary for generating reliable evidence. This would aid understanding of how study characteristics change over time and across therapeutic areas, as well as the progress of initiatives to improve the quality and relevance of clinical studies.11–15 Therefore, we approached the UK Health Research Authority (HRA), which oversees national ethical review of health research conducted within the National Health Service (NHS). With the support of the HRA and having signed the appropriate confidentiality declarations, we developed a pilot project to determine the feasibility of interrogating with minimal manual intervention the data contained in HARP (HRA Assessment Review Portal). HARP is a web-based management information system used by the HRA for all research ethics applications that require NHS REC (National Health Service Research Ethics Committee) review and/or HRA approval. We negotiated access to data on clinical studies submitted in 2015. The extract had personal data and organisational identifiers removed by HRA and did not include phase I healthy volunteer studies. Our pilot focused on automatically summarising the characteristics of clinical studies, including therapeutic area, blinding, randomisation, use of Independent Data Monitoring Committee (IDMC), patient and public involvement (PPI) in the research, dissemination of the study results and use of new methodologies such as adaptive designs.

Methods

The HRA collects applications for REC ‘approval’ (a favourable opinion) through the Integrated Research Application System (IRAS)16 and stores them in HARP. IRAS is a web-based system used to capture the information required by review bodies in the UK, including the Medicines and Healthcare Products Regulatory Agency and HRA. When the forms are completed, IRAS saves the results as an XML document. XML is a commonly used file format that stores structured data in plain text. The XML document can be used to generate PDF copies of the required forms and to repopulate the online web tool so users can later edit their applications. Once the appropriate permissions were agreed, the HRA provided us with anonymised applications for clinical studies submitted for review from 2014 to 2016, excluding phase I healthy volunteer studies. For this pilot, we wished to focus on the extraction of clinical study characteristics of interest from the PDF versions of the HRA forms underlying the IRAS application system. We were provided with an incomplete mapping between the questions on the PDF and the data stored in the XML document, such that many questions relevant to our study were not immediately labelled; nor was it apparent which questions were effectively switched on and off for applicants in completing the system’s initial filter questions for their application. Therefore, we reverse-engineered the structure of the data in IRAS to determine which XML tags were responsible for creating the content we wanted to extract from the XML files to study. The IRAS tool has been updated over time; not all of applications were created using the same version. Although the majority (91%) of the XML documents were created with one version (the latest version of the tool in the dataset), eight different versions were present in the dataset we received. We were not provided with the documentation detailing the ways in which the versions differed; we had access only to the latest version. Our inspection of the data indicated that the differences relevant to this pilot study were minor and we corrected for these when they were detected; however, there may be some undetected differences where the distinction between versions was important for our study. For our pilot project, we chose to collect information on: the number of randomised controlled trials (RCT); the trial phase; the therapeutic area; the number of clinical trials of investigational medicinal products (CTIMPs — trials that require registration with regulatory body); the research method used (eg, RCT, feasibility/pilot study, blinding); the use of blinding; the use of systematic or formal literature reviews in planning the study; the use of adaptive or Bayesian designs; the use of a data monitoring committee; the plans for dissemination of findings; and the role of PPI in aspects of the research. Success would be measured in terms of our ability to access the information and generate descriptive summaries with minimal need for manual data reviews. The data for our target areas were either held in multiple choice questions that we could identify and summarise or in free-text fields from which information needed to be determined and extracted. For those questions, we preprocessed the text using Spacy (https://spacy.io/), a natural language processing (NLP) package for Python. This involved identifying sentence boundaries and performing morphological analysis to convert words into their canonical dictionary form. We then used regular expressions to match phrases we were interested in, ignoring differences relating to spacing, hyphenation and capitalisation; Roman numerals were standardised to Latin numbers. For example, when looking for evidence that a systematic review was either performed or an existing review used during the study planning process, we searched for phrases such as ‘systematic review’, ‘reviewed systematically’, ‘literature review’, ‘evidence review’ and ‘evidence-based review’ in the answers provided to the relevant questions on the form, or instances where these words were used near each other, where ‘near’ was defined as being within three words. For adaptive designs, we employed the US Food and Drug Administration (FDA) definition, namely ‘a study that includes a prospectively planned opportunity for modification’.17 We extracted the sentence containing the target phrase as well as one sentence before and after into a separate document. These textural extracts were then reviewed by the authors to identify true and false positives. The data extracted from the separate XML files were collated into one dataset and descriptive analyses of this dataset were done with Stata.

Results

We received (30 June 2016) the XML files for 1814 application records submitted for ethical review during the period specified, and extracted by HRA from the IRAS system. Of these, 1659 (92%) were from 2015, 154 (9%) from 2014 and 1 was from 2016. Three records were corrupted and could not be processed. In discussion with HRA, we discarded some records to remove duplicates. We kept one of four entries for the WHEAT trial which had been sent to multiple committees as part of research on consistency of REC opinions.18 We discarded one of two entries for a further study, which was initially given a favourable opinion and then submitted additionally to a specialist REC in Scotland (again, favourably). Thirty-two further records were discarded that initially had an unfavourable opinion before re-submission as a (near) identical, separate application for review. We discarded the first applications and kept the re-submissions, regardless of the subsequent review’s outcome. Therefore, our dataset included 1775 studies. The filter questions switched off, for some applications, questions in which we were interested. For example, trial phase was only recorded for CTIMPs and the use of data monitoring committees was infrequently recorded. Furthermore, the use of blinding was not sufficiently well captured in the system to allow us to present reliable data. A total of 963/1775 (54%) of the applications were stated as being RCTs. Table 1 describes the disease setting, research method, CTIMP and REC opinion by whether they were an RCT. The most common research area, using the categorisation of the IRAS system, was cancer, followed by cardiovascular, musculoskeletal and respiratory diseases; around one-third of records specified more than one area. This is broadly consistent with previous reviews.19 20 Around one-fifth of RCTs were pilot studies; one-third of records were employing more than one method. The REC gave a favourable opinion to the vast majority of applications (1740/1775, 98%), but required additional information for most of these 1323/1740 (75%).
Table 1

Overview of research activities submitted for REC approval

CharacteristicRCTOtherTotal
N%N%N%
N9638121775
Therapeutic area*
 Cancer168171922436020
 Cardiovascular9410941218811
 Musculoskeletal97103041629
 Respiratory97103041599
 Paediatrics5663041488
 Neurological7383041458
 Mental health6673041177
 Inflammatory and immune system7273041136
 Oral and gastrointestinal6473041136
 Blood4456781116
 Diabetes586385965
 Infection556304925
 Renal and urogenital394304825
 Generic health relevance465304764
 Metabolic and endocrine323304704
 Dementias and neurodegenerative273364634
 Skin364304573
 Reproductive health and childbirth303304563
 Stroke273304533
 Eye242283523
 Injuries and accidents222304362
 Congenital disorders101202302
 Ear6171131
Count of therapeutic areas claimed
 None5254961016
 16606949561115565
 2201212002540123
 3293496784
 ≥4212192402
Research methods used*
 RCT9631002096554
 Feasibility/pilot study177183003747727
 Questionnaire98101391723713
 Cohort observation2321612018410
 Qualitative research9097691669
 Controlled trial, no randomisation30157191609
 Laboratory study303476774
 Case series/case note review71456523
 Cross-sectional study40314352
 Case–control study51253302
 Database analysis81172251
 Epidemiology7151121
 Meta-analysis000000
 Other6872693333719
Count of research methods used
 164167%52264116366
 2192201602035220
 3738971217010
 4485253734
 57171141
 6201030
CTIMP
 Unlicensed315331892350428
 New use1461570921612
 Within SmPC667263925
 Other10151151
Any of these 515 53 284 35
Involves ionising radiation
 Yes6666958272124871
 No297312222851929
Missing 0NA8NA8NA
REC opinion
 Favourable9429879898174098
  Favourable200212172741723
  Favourable (extra info)7427758172132375
 Unfavourable212142352

*Not mutually exclusive.

CTIMP, clinical trials of investigational medicinal product; RCT, randomised controlled trial; REC, Research Ethics Committee; SmPC, summary of product characteristics.

Overview of research activities submitted for REC approval *Not mutually exclusive. CTIMP, clinical trials of investigational medicinal product; RCT, randomised controlled trial; REC, Research Ethics Committee; SmPC, summary of product characteristics. Over half of the RCTs (515/963, 53%) were CTIMPs, of which 315/515 (61%) were CTIMPs of unlicensed products, 146/515 (28%) licensed CTIMPs in a new setting and 66/515 (13%) CTIMPs used according to their summary of product characteristics (SmPC); the categories were not mutually exclusive. Over a third (284/812, 35%) of the other studies (not RCTs) were also CTIMPs. Table 2 describes the plans in design and dissemination. Most RCTs (895/963, 93%) were not recorded as being preceded by a formal systematic review of the literature. Only 15 (2%) RCTs were detectably designed under a Bayesian framework; 20 (2%) were detectable as having an adaptive design. There were 34 cluster randomised and 5 stepped wedge trials, overall. Six hundred fourteen (63%) RCTs were requesting new biological samples; 120 (12%) RCTs were seeking access to previous biological samples.
Table 2

Issues in design and dissemination in all entries

CharacteristicRCTOtherTotal
N%N%N%
N9638121775
Review of data as part of development
 Neither8959375993165493
 Systematic review only384324704
 Meta-analysis only182132312
 Both12181201
Design characteristics
 Neither926967789617041
 Adaptive design20291292
 Bayes design152243392
 Both201030
 Cluster randomised26381342
 Stepped wedge302050
Sample collection*
 Taking new samples6146442252103658
 Accessing stored samples120121101423013
Plans for dissemination†
 Peer-reviewed scientific journal9129575593166794
 Conference presentation869907168158589
 Internal report6416752665116766
 Publication on website536563674590351
 Submission to regulatory authority475493524382747
 Other publication239251782241723
 Access to raw data19120133161146
 Other003032418
 No plans to report or disseminate30

*Not asked for non-regulatory RCTs.

†Not mutually exclusive.

RCT, randomised controlled trial.

Issues in design and dissemination in all entries *Not asked for non-regulatory RCTs. †Not mutually exclusive. RCT, randomised controlled trial. The large majority of research (1667/1775, 94%) was planned for dissemination in the peer-reviewed literature and most (1585/1775, 89%) were planned for conference presentations. A total of 475/963 (49%) RCTs and 352/812 (43%) other studies were planned for submission to regulatory authorities. A small number, 114/1775 (6%), were planning to offer raw data to external applicants as a key form of dissemination. Table 3 notes the reported patients and public involvement (PPI) using the categories specified by IRAS. A total of 726/1775 (41%) of studies claim PPI in the design and 659/1775 (37%) claim planned PPI in dissemination of the findings. PPI engagement in undertaking and management of the research was less common, and few studies (147/1775, 8%) involved planned PPI in analysis. Around one-third of studies involve PPI in two or more of these IRAS-defined areas.
Table 3

Reported PPI

CharacteristicRCTOtherTotal
N%N%N%
N9638121775
Areas of PPI activity*
 Design of the research403423234072641
 Dissemination of findings375392843565937
 Undertaking the research242252192746126
 Management of the research225231291635420
 Analysis of results7277591478
Number of areas of PPI activity
 None389403023769139
 1205212092641423
 2139141171425614
 312513931121812
 46673751036
 All394324714

*Not mutually exclusive.

PPI, patient and public involvement.

Reported PPI *Not mutually exclusive. PPI, patient and public involvement. Table 4 shows within CTIMPs the similarities and differences across the phase of research. Trial phase was asked as a series of separate yes/no questions for phase I, II, III and IV. These were not mutually exclusive. For the purpose of summarising, we selected the highest level if more than one option was selected. Early phase trials were more likely to be designed within a Bayesian framework; phase III trials most often reported the use of a placebo and the use of an Independent Data Monitoring Committee; phase IV trials were least likely to be submitted to a regulatory authority or put on a website, but most often reported PPI in all five broad areas.
Table 4

Characteristics of the CTIMPs by trial phase

Trial phase*Ph IPh IIPh IIIPh IVTotal
CharacteristicN%N%N%N%N%
All entries, N9928732498808
Design characteristics
 Neither86892619330394909474093
 Adaptive117213433243
 Bayes10101245222294
 Both0021001130
 Cluster randomised0000000000
 Stepped wedge0000000000
 Involves placebo21221234416351252633242
Independent data monitoring committee
 No408284596142467823158
 Yes91859418458132216542
Missing 8 139 176 37 400
Methods of dissemination
 Peer-reviewed scientific journal8689265942909093978689
 Submission to regulatory authority8588215762728548508588
 Conference presentation8285253902718489938285
 Internal report7173212752598160637173
 Publication on website5658161572036345475658
 Other publication26277627922917182627
 Access to raw data881242588888
 Other12124717481518191212
 No plans to report or dissemination0000210000
Areas of PPI activity†
 Design of the research26278229391250522627
 Management of the research1010341228931321010
 Undertaking the research14144717361129301414
 Analysis of results1115572121311
 Dissemination of findings20217426642044462021
 None of the above53551495322369353646058
REC opinion
 Favourable97982779731898939578597
  Favourable141451184012151512015
  Favourable (after extra info)83842267927886788066582
 Unfavourable00523133111

*Trial phase calculated.

†Not mutually exclusive.

CTIMP, clinical trials of investigational medicinal product; PPI, patient and public involvement; REC, Research Ethics Committee.

Characteristics of the CTIMPs by trial phase *Trial phase calculated. †Not mutually exclusive. CTIMP, clinical trials of investigational medicinal product; PPI, patient and public involvement; REC, Research Ethics Committee. Table 5 shows the outcome of the manual review of the extracted free-text fields. Accuracy, defined as number of true positives divided by overall total ranged from 43% for adaptive designs to 100% for stepped wedge. The false positives related to studies described as ‘phase 1/2’ or ‘phase 2/3’, but with no evidence of an adaptive step, for example, phase 1/2 was often used to describe a classical pharmacokinetic study in patients rather than healthy volunteers; the database did not contain healthy volunteer studies. For ‘systematic review’ and ‘meta-analysis’, the search often picked up references to previous studies and not the planned study.
Table 5

Review of possible search terms

Design element Search termPerformance of search term*
At sentence levelAt study level
True positiveFalse positiveAccuracyTrue positiveFalse positiveAccuracy
Systematic review
 Review (near) systematically2010020100
 Evidence (near) review41804180
 Systematic (near) review983972582372
 Literature (near) review291369261268
Evidence-based review 0 0 NA 0 0 NA
 Summary*†1335372903671
Meta-analysis
 Pooled analysis2010020100
 Meta-analysis773171491972
Integrated analysis 0 0 NA 0 0 NA
 Summary*†793172511973
Adaptive design
 Adaptive design190100130100
 Adaptive randomisation8010050100
 Continual reassessment2010010100
 Sample size re-estimation15010090100
 Seamless design1010010100
 Phase I/II44393378
 Phase II/III22572238
CRM design 0 0 NA 0 0 NA
Drop-the-loser 0 0 NA 0 0 NA
MAMS design 0 0 NA 0 0 NA
Multiarm multistage 0 0 NA 0 0 NA
Pick-the-winner 0 0 NA 0 0 NA
 Summary*†516843316034
Bayes
 Bayes8329842198
 Summary*†8329842198
Cluster RCT
 Cluster RCT260100110100
 Control cluster6010040100
 Randomised cluster2010010100
 Cluster random4619830197
 Summary*†8019934197
Stepped wedge
 Step (near) wedge1010010100
 Stepped (near) wedge16010050100
 Summary*†17010050100

NB The table reports on whether each search term accurately identified a specific design element of the study in selected free-text fields. All variations of the spelling, capitalisation, hyphenation and so on were covered in the search terms used.

*The left side of the table separately considers each sentence which matched a search term accurately identified the design property. There were 613 such sentences. The Summary row in each subsection is the sum of the constituent search terms.

†The right side of the table separately considers whether any sentence in a study matching a search term accurately identified the design property. Some studies matched a search term repeatedly, therefore the numbers are smaller. The Summary row in each subsection reflects a search across all of its constituent search terms. Note that some studies matched more than one search term in a subsection, so the Summary row may not be the sum of the rows above for example, a study that uses both ‘cluster random’ and ‘control cluster’ would appear only once in the Summary row, and would appear separately in the ‘cluster random’ and ‘control cluster’ rows.

CRM, continual reassessment method; MAMS, multi-arm multi-stage; RCT, randomised controlled trial.

Review of possible search terms NB The table reports on whether each search term accurately identified a specific design element of the study in selected free-text fields. All variations of the spelling, capitalisation, hyphenation and so on were covered in the search terms used. *The left side of the table separately considers each sentence which matched a search term accurately identified the design property. There were 613 such sentences. The Summary row in each subsection is the sum of the constituent search terms. †The right side of the table separately considers whether any sentence in a study matching a search term accurately identified the design property. Some studies matched a search term repeatedly, therefore the numbers are smaller. The Summary row in each subsection reflects a search across all of its constituent search terms. Note that some studies matched more than one search term in a subsection, so the Summary row may not be the sum of the rows above for example, a study that uses both ‘cluster random’ and ‘control cluster’ would appear only once in the Summary row, and would appear separately in the ‘cluster random’ and ‘control cluster’ rows. CRM, continual reassessment method; MAMS, multi-arm multi-stage; RCT, randomised controlled trial.

Discussion

We achieved our primary feasibility aim of negotiating access to the centralised UK approvals system and devising a way of extracting information from a series of separate XML files. We anticipate that the programmes we developed could extract further pertinent points of information with minimal manual involvement. We were also able to achieve our primary descriptive aim of systematically reporting on the state of clinical research in the UK over around 1 year, focused on 2015. We found that nearly 1000 RCTs were submitted for approval in the UK during that period and >800 other studies. A key message is the volume of research activity in the UK, with the country demonstrably research active. The majority of applications received a favourable opinion but most required further information first, suggesting that applications could be better completed, saving time and effort for all parties. It was notable that the reported use of adaptive designs was low and was not markedly different from earlier estimates.19 20 This is disappointing, as many members of this design family have scope to reduce time to answer and/or reduce the average cost per answer, particularly by moving away early from treatment approaches that are not likely to improve outcomes sufficiently for patients.21 However, Bayesian approaches have penetrated 10% of early phase CTIMPs. The reported use of IDMCs for phase III CTIMPs was, perhaps, low at 58%, but this is a substantial increase over the prevalence in trials published in high impact-factor journals in 1990 and 2000.22 Involvement of patients and the public in various aspects of research is increasingly seen as important, but reported rates of engagement were quite low: more than half of studies had no PPI involvement.23 Cursory review of the free-text fields associated with these PPI categories suggests that some applicants may have been a little generous in choosing to select a particular category, suggesting that the numbers we report may be overestimates. Other comments suggest that PPI is not needed, perhaps reflecting that the value of PPI is yet to be recognised by some researchers.3 Further research in this area is required. Most but not all studies were planned for publication; there is still some way to go on this transparency aim. That raw data would be made available for some studies is encouraging and we hope that this will increase with time for appropriate projects from qualified researchers.24 Extracting the information from the HRA system was more effortful than expected. Moreover, important characteristics of clinical trials like blinding could not be analysed and certain search terms clearly had limitations. The issue of blinding is particularly important, as it is one of the desirable characteristics in generating reliable evidence from clinical trials.11 However, now that we have completed this feasibility step, it is possible for future research projects to access an expanded dataset and work with the HRA to organise the database in a way that will facilitate its analysis. Access to several years of data would allow researchers, as well as the HRA itself, to examine the progress of initiatives to improve the quality of clinical research, such as those designed to encourage PPI and publication of clinical trial results.13 15 We would also encourage the HRA to adopt standard definitions to describe the characteristics of clinical trials, as done in the USA.25 Confusion over the definition of terms greatly complicates efforts to automatically analyse clinical research. Our pilot work has some limitations. We had no straightforward way of aligning the questions in the PDF document with the information stored in the XML document. Perhaps if we had negotiated access to the backend software used by IRAS to generate these PDF forms, we may have been able to do this automatically; instead, we needed to reverse-engineer the IRAS system in order to figure out which XML tags were responsible for creating the content we wanted to study. Not all of the applications had been created using the same version of the IRAS tool.16 Although a large majority of the documents were created using the latest version of the tool, eight different versions were present in the dataset we received. We did not receive any guidance on the ways in which one version of the tool differed from another version, and we only had access to the latest version. We made efforts to determine the specific situations in which questions were not asked rather than asked but not answered, in order to make this distinction ourselves, though ideally this could be indicated directly in the XML dataset. The labelling of missing answers and inactivated questions was not consistent. There is scope for applicants to misunderstand the questions and misrepresent the data. For example, three trials were noted as using RCT methodology but answered ‘no’ to the question: ‘Will participants be allocated to groups at random?’ We checked the data and found these were not actually RCTs; we included them with the ‘other studies’. Likely this error by the applicants activated or de-activated some questions inappropriately. In this example, the RCT method question actually asked about randomisation of individuals and these three applicants who submitted cluster randomised trials had answered this question inconsistently. Finally, the search terms chosen for the free-text fields were not exhaustive and because it was unfeasible in a pilot to review all applications, the sensitivity and specificity of the free-text extraction could not be calculated. Some interesting information was held only in the free-text fields. We used standard methods to extract this information using automation supplemented by manual checking. There is a wealth of free-text information, which through the application of text-mining techniques could provide incredibly valuable insights into the characteristics of health research. For this pilot study, we did not have the resources to review all of the questions and we particularly could not spend time reading free-text fields. Therefore, we felt obliged not to correct categorical questions if we noticed they were contrasted by free-text; however, a fully-resourced study could do this. For example, we noticed at least one entry, which ticked boxes for PPI engagement and expanded on this by claiming that, as doctors, they knew what patients wanted and did not need to trouble patients for their time on these aspects. In conclusion, we have demonstrated that anonymised data from the HRA’s research system are accessible and can be queried for information. We strongly encourage the development of fully resourced collaborative projects to delve more deeply into this data. We believe that it is imperative that the characteristics of clinical research in the UK are understood, as these underpin clinical guidance. Finally, there are many ongoing initiatives to improve the quality of clinical research, but only by fully understanding the state of, and changes to, the research profile of the UK, can we appreciate their impact.
  16 in total

1.  Sharing clinical trial data: a proposal from the International Committee of Medical Journal Editors.

Authors:  Darren B Taichman; Joyce Backus; Christopher Baethge; Howard Bauchner; Peter W de Leeuw; Jeffrey M Drazen; John Fletcher; Frank A Frizelle; Trish Groves; Abraham Haileamlak; Astrid James; Christine Laine; Larry Peiperl; Anja Pinborg; Peush Sahni; Sinan Wu
Journal:  Lancet       Date:  2016-01-23       Impact factor: 79.321

2.  Reported use of data monitoring committees in the main published reports of randomized controlled trials: a cross-sectional study.

Authors:  Matthew R Sydes; Douglas G Altman; Abdel B Babiker; Mahesh K B Parmar; David J Spiegelhalter
Journal:  Clin Trials       Date:  2004-02       Impact factor: 2.486

3.  Characteristics of clinical trials registered in ClinicalTrials.gov, 2007-2010.

Authors:  Robert M Califf; Deborah A Zarin; Judith M Kramer; Rachel E Sherman; Laura H Aberle; Asba Tasneem
Journal:  JAMA       Date:  2012-05-02       Impact factor: 56.272

Review 4.  The Clinical Trials Transformation Initiative (CTTI).

Authors:  Alberto Grignolo
Journal:  Ann Ist Super Sanita       Date:  2011       Impact factor: 1.663

5.  All trials must be registered and the results published.

Authors:  Iain Chalmers; Paul Glasziou; Fiona Godlee
Journal:  BMJ       Date:  2013-01-09

6.  SPIRIT 2013: new guidance for content of clinical trial protocols.

Authors:  An-Wen Chan; Jennifer M Tetzlaff; Douglas G Altman; Kay Dickersin; David Moher
Journal:  Lancet       Date:  2013-01-08       Impact factor: 79.321

Review 7.  Patient involvement in clinical research: why, when, and how.

Authors:  José A Sacristán; Alfonso Aguarón; Cristina Avendaño-Solá; Pilar Garrido; Juan Carrión; Alipio Gutiérrez; Robert Kroes; Angeles Flores
Journal:  Patient Prefer Adherence       Date:  2016-04-27       Impact factor: 2.711

8.  Research ethics committee decision-making in relation to an efficient neonatal trial.

Authors:  C Gale; M J Hyde; N Modi
Journal:  Arch Dis Child Fetal Neonatal Ed       Date:  2016-09-14       Impact factor: 5.747

9.  Discrepancies in sample size calculations and data analyses reported in randomised trials: comparison of publications with protocols.

Authors:  An-Wen Chan; Asbjørn Hróbjartsson; Karsten J Jørgensen; Peter C Gøtzsche; Douglas G Altman
Journal:  BMJ       Date:  2008-12-04

10.  Speeding up the evaluation of new agents in cancer.

Authors:  Mahesh K B Parmar; Friederike M-S Barthel; Matthew Sydes; Ruth Langley; Rick Kaplan; Elizabeth Eisenhauer; Mark Brady; Nicholas James; Michael A Bookman; Ann-Marie Swart; Wendi Qian; Patrick Royston
Journal:  J Natl Cancer Inst       Date:  2008-08-26       Impact factor: 13.506

View more
  1 in total

1.  Access to routinely collected health data for clinical trials - review of successful data requests to UK registries.

Authors:  Sarah Lensen; Archie Macnair; Sharon B Love; Victoria Yorke-Edwards; Nurulamin M Noor; Meredith Martyn; Alexandra Blenkinsop; Carlos Diaz-Montana; Graham Powell; Elizabeth Williamson; James Carpenter; Matthew R Sydes
Journal:  Trials       Date:  2020-05-12       Impact factor: 2.279

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.