| Literature DB >> 30873409 |
Angelo Silverio1, Pierpaolo Cavallo2, Roberta De Rosa1, Gennaro Galasso1.
Abstract
Cardiovascular disease (CVD) accounts for the majority of death and hospitalization, health care expenditures and loss of productivity in developed country. CVD research, thus, plays a key role for improving patients' outcomes as well as for the sustainability of health systems. The increasing costs and complexity of modern medicine along with the fragmentation in healthcare organizations interfere with improving quality care and represent a missed opportunity for research. The advancement in diagnosis, therapy and prognostic evaluation of patients with CVD, indeed, is frustrated by limited data access to selected small patient populations, not standardized nor computable definition of disease and lack of approved relevant patient-centered outcomes. These critical issues results in a deep mismatch between randomized controlled trials and real-world setting, heterogeneity in treatment response and wide inter-individual variation in prognosis. Big data approach combines millions of people's electronic health records (EHR) from different resources and provides a new methodology expanding data collection in three direction: high volume, wide variety and extreme acquisition speed. Large population studies based on EHR holds much promise due to low costs, diminished study participant burden, and reduced selection bias, thus offering an alternative to traditional ascertainment through biomedical screening and tracing processes. By merging and harmonizing large data sets, the researchers aspire to build algorithms that allow targeted and personalized CVD treatments. In current paper, we provide a critical review of big health data for cardiovascular research, focusing on the opportunities of this largely free data analytics and the challenges in its realization.Entities:
Keywords: acute coronary syndromes; big data; cardiovascular disease; coronary artery disease; electronic health records; heart failure
Year: 2019 PMID: 30873409 PMCID: PMC6401640 DOI: 10.3389/fmed.2019.00036
Source DB: PubMed Journal: Front Med (Lausanne) ISSN: 2296-858X
EHR resources relevant for cardiovascular research.
| CArdiovascular research using LInked Bespoke studies and Electronic health Records (CALIBER) | UK | 10,000,000 subjects | |
| Mondriaan | NE | 15,000,000 subjects | |
| ABUCASIS | ES | 5,000,000 subjects | |
| National Institute for Health Research Health Informatics Collaborative (HIC) | UK | ||
| SWEDEHEART | SE | 2,000,000 subjects | |
| European Society of Cardiology European Research Programme (EORP) | EU | 2,200 centers 100,000 subjects | |
| National Institute for Cardiovascular Outcomes Research (NICOR) | UK | 200,000 records on HF 1,000,000 records on CRD 450,000 records on PCI | |
| UK-Biobank | UK | 500,000 subjects | |
| European Prospective Investigation into Cancer and Nutrition (EPIC) -CVD | EU | 10 countries 500,000 subjects | |
| China Kadoorie Biobank | China | 500,000 subjects | |
| UCL-LSHTM-Edinburgh-Bristol (UCLEB) Consortium | UK | 30,000 subjects | |
| US Department of Veteran Affairs-Million Veteran Program | US | 500,000 subjects | |
| Kaiser Permanente-Research Program on Genes, Environment and Health | US | 500,000 subjects | |
| eMERGE | US | 105,000 subjects | |
| DiscovEHR project of the Regeneron Genetics Center and the Geisinger Health System | US | 42,000 subjects | |
| Precision Medicine Initiative Cohort Program | US | ||
| Vanderbilt BioVU | US | ||
| GENIUS-CHD | Global | 250,000 subjects | |
| HERMES Consortium | Global | 30,000 subjects | |
| AFGen Consortium | EU-US | ||
CRD, cardiac rhythm diseases; ES, Spain; EU, Europe; HF, heart failure; NE, Netherlands; PCI, percutaneous coronary intervention; SE, Sweden; UK, United Kingdom; US, United States.
Main opportunities and challenges of EHR for cardiovascular research.
| •High-resolution large-scale studies | Large cohorts allow to study infrequent events |
| •Public health improvement | EHR improve quality of healthcare and spending control |
| •Development of predictive models | Machine learning predictive models do not require statistical assumption and use complex algorithm for the analysis of large and heterogeneous dataset |
| •Timely answers to cardiology controversy | Big data provide real-time response to the problems of daily clinical practice |
| •Drug surveillance | EHR-based post-marketing surveillance of adverse drug events |
| •International comparison | Assessment of performance of healthcare system of different countries in term of patients' outcome and costs |
| •Integrating pharmacogenomics | Informatics models by disseminating patient information at the point of care may facilitate the development of pharmacogenomic clinical decision support in daily practice |
| •Personalized medicine by estimates of benefits and harms of treatments | |
| •Quality of care and performance measures | Monitoring quality of treatments and support continuous improvement in the participating centers |
| •Drug repurposing | Data linkage in big dataset may help to identify new uses for approved or investigational drugs that are outside the scope of the original medical indication |
| •Genetic insights | |
| •Disease definition | Heterogeneous and not standardized disease definitions are a challenge for computation |
| •Source availability | |
| •Data sharing | The practice of making data used for research available to other investigators. |
| •Data quality and missing data | |
| •Translational applicability of results | Apply findings from big data analytics to enhance diagnosis, treatment and prognostic stratification of diseases |
| •Dependence problem | Situation in which a program instruction is dependent on a result of a sequentially previous instruction before it can complete its execution |
| •Data linkage | Method of bringing together informations from different sources about the same person or clinical entity |
| •Data inconsistency | If the same data is stored in different formats in two files and matching of data must be done between files. Moreover, these files duplicate some of the data |
| •Interpretation of results | |
| •Unstructured data | Processing data not having a pre-defined model or not organized in a pre-defined manner |
| •Data integrity | Data integrity is the maintenance of the accuracy and consistency of data over its entire life-cycle |
| •Training | |
| •Legal and ethical issues | |
| •Data security | |
EHR, electronic health records.
Figure 1Big health data overview: from sources to potential clinical application. AP, arterial pressure; CT, computed tomography; ECG, electrocardiography; HR, heart rate; ICD, implantable cardioverter defibrillator; MRI, magnetic resonance imaging; OTC, over the counter; SaO2, arterial oxygen saturation; SNPs, single nucleotide polymorphisms.