Literature DB >> 28093878

In Pursuit of Greater Reproducibility and Credibility of Early Clinical Biomarker Research.

L M McShane1.   

Abstract

Entities:  

Mesh:

Substances:

Year:  2017        PMID: 28093878      PMCID: PMC5355975          DOI: 10.1111/cts.12449

Source DB:  PubMed          Journal:  Clin Transl Sci        ISSN: 1752-8054            Impact factor:   4.689


× No keyword cloud information.

INTRODUCTION

Biomarkers underlie many clinical tests that are integral to the practice of personalized medicine. Reproducibility and scientific credibility of clinical biomarker early development studies are critical to avoid advancing worthless or potentially harmful biomarker‐based tests into late‐phase clinical studies and clinical practice. This commentary discusses key aspects to consider when conducting and evaluating early clinical biomarker research. Greater attention to these aspects would enhance research reproducibility and better prioritize biomarkers for further clinical development. Recognition of the problem of irreproducibility of preclinical drug development research led to a call for transparent reporting standards and recommendations for improved study designs.1 Similar principles apply to early research aiming to develop clinical biomarker tests (henceforth termed “early clinical biomarker research”), but there are important differences too. A major difference between preclinical drug development studies and early clinical biomarker development studies is that the latter are often conducted retrospectively using stored specimens collected in routine clinical care settings or in the context of research studies originally addressing different questions. Thus, early clinical biomarker research has features of retrospective observational studies that extend beyond the experimentally controlled settings typical for preclinical drug development research. The development process for biomarker‐based tests usually begins with a study aiming to establish whether a biomarker is associated with some clinical outcome or other phenotype. The test may be based on a single biomarker or a panel of biomarkers combined via a statistical prediction model; for example, using “omics” assay technologies that measure “related sets of biological molecules in a comprehensive fashion.”2 Further development requires a series of studies to gather more evidence and eventually incorporate the biomarker into a clinical test that is validated for a specific clinical use. The clinical role for a biomarker‐based test typically falls into one or more of the following categories (see US Food and Drug Administration / National Institutes of Health glossary at https://www.ncbi.nlm.nih.gov/books/NBK326791/): diagnostic, monitoring, pharmacodynamics/response, predictive, prognostic, safety, and susceptibility/risk. This commentary focuses on overarching principles to consider in early clinical biomarker research to enhance reproducibility and provide a solid foundation for later stages of development. For more extensive discussion of best practices to be applied throughout the clinical biomarker development process, readers are referred elsewhere.2, 3, 4

Study design and primary data generation

Study design is usually discussed in the context of prospectively conducted preclinical experiments or clinical trials, but many design principles apply also for studies with retrospective elements. Attention should focus on biomarkers that have potential to provide insights into biological processes or translate into tools for clinical decision‐making. Building on that foundation, good study design requires recognition of the many factors that can lead to variation in results (systematic and random). Biomarker assay methods and subject and specimen factors may systematically affect biomarker measurements and their associations with clinical outcomes. In early development studies, biomarker assays should meet at least minimal analytical performance standards to establish that the assay measures the intended analyte and has acceptable reproducibility over the range of values relevant to the clinical setting. Performance criteria become more stringent as the development proceeds (see Supplementary Table S1). Assay methods should be documented carefully to facilitate replication. Subject factors such as age or gender; disease status, subtype or stage; and comorbidities should be considered in formulating retrospective eligibly criteria. Requirements for specimen collection, processing, and handling to ensure reliable assay performance should be defined. Designs confusing important subject‐ or specimen‐related factors with biomarker or outcome status must be diligently avoided. For example, women would be inappropriate control subjects in a study of prostate cancer detection biomarkers. Multi‐institutional studies increase the risk of biases. Differences in patient characteristics, clinical management, and specimen handling across institutions may confound associations between biomarker measurements and outcomes. Preferably, studies are designed to avoid such confounding, or at minimum, information should be collected to attempt adjustment for these factors in analyses. Many poorly designed studies exist in the published literature, and their reproducibility is often compromised. Random factors are those that differ from study to study and generally cannot be controlled completely. Examples include laboratory assay batches and observers recording subjective biomarker or outcome assessments. “Omics” assays generating large numbers of measurements per specimen are particularly prone to batch effects due to their sensitivity to subtle changes in laboratory conditions.5 Effects of random (or not easily standardized) factors can be reduced through randomization and blinding. An example of poor study design is running samples from subjects with a favorable disease outcome in one assay batch and samples from subjects with an unfavorable outcome in another. Samples should be randomized to assay batches or allocated in a way that batch effects could be eliminated through statistical corrections. Observers recording subjective clinical outcomes should not be confounded with biomarker status and should remain blinded to biomarker values. Similarly, insidious biases can occur when individuals making subjective biomarker assessments are not blinded to subjects’ clinical outcomes. Inattention to these design issues can impair study reproducibility. Sample size (number of study subjects) is another important study design consideration. It may be based on calculated power for a statistical test or precision for an estimate of a parameter of interest. Example parameters are accuracy of a biomarker in identifying individuals who respond to (or experience toxicity from) a drug, or a hazard ratio representing a biomarker's prognostic association with clinical outcome. Such calculations help to set expectations for evidence to be gained but should be performed prior to study initiation and based on realistic assumptions. Related to sample size is within‐subject replication (number and types of replicate measurements per subject). When measurement error for outcome or other variables is substantial, replication can reduce noise. Example replicate types include biomarker measurements on samples collected over several timepoints and repeated measurements on a single sample. Measurement replication schemes may be important to mimic when attempting to reproduce study results. Total number of observations must not be confused with number of independent subjects, and data analyses must account appropriately for within‐subject replication schemes.

Data collection and curation

There are rigorous quality standards for collection and curation of clinical trial data. In contrast, many early biomarker studies rely on clinical characteristics and outcome data collected retrospectively, possibly extracted from clinical charts, registries, or electronic medical record systems. Investigators should make efforts to confirm the validity of such retrospectively collected data to ensure that they are accurate and correctly interpreted. Data from these sources together with newly generated biomarker data also need to be managed with care. Risk of inadvertent data corruption is increased with inexperienced or careless use of software with sorting, cut‐and‐paste, and autocorrect features. Omics data present additional challenges due to their sheer volume and specialized formats, which require complex data systems managed by experienced personnel.

Data analysis

Many early biomarker investigations are conducted without a statistical analysis plan prespecifying primary analyses or details of analysis approaches. The number of analyses can easily reach dozens considering different end points, subgroups, explanatory variables and models, or cutpoints applied to continuous biomarker values. Chances of false‐positive findings increase, as each additional analysis may generate false‐positive findings from noise in the data. Pitfalls of conducting numerous exploratory analyses are well recognized by clinical trial methodologists.6 Outlining key analyses in a prespecified analysis plan helps to distinguish preplanned analyses from data‐driven exploratory or ad hoc analyses that are more likely to generate false‐positive or biased results. Data analysis approaches must be consistent with study design, including accounting for nonrandom selection of study subjects. Use of case–control and matched study designs is fairly common in retrospective biomarker studies, and these require specialized statistical analysis methods.7, 8 Analyses should additionally account for multiple testing, data distributions (e.g., nonnormal data), functional relationships between biomarkers and outcomes (e.g., nonlinear), correlations between multiple measurements per study subject, and handling of outliers and missing data. Statistical analyses cannot rescue data that are corrupted or generated by terribly flawed study designs; in the opposite direction, inappropriate statistical analyses can lead to misleading results and inappropriate conclusions even when based on high‐quality data.

Results interpretation and study reporting

Complete and transparent reporting of study design, conduct, analysis, and results facilitates proper interpretation of a study and evaluation of its quality. Others may be unable to reproduce results of a study if not adequately informed about the study population, specimen requirements, and biomarker assay methodology. Different data analysis approaches may lead to different results, so it is important to describe analyses that were performed and why those approaches were selected. Disclosure of the total number of analyses performed and which were prespecified is important to gauge potential for false‐positive findings. Study sample size and precision of estimated effects or parameters of interest should be reported to indicate the strength of evidence; for example, to help distinguish nonsignificant from convincingly null findings. Relevant parameters to report will differ depending on the potential clinical role for the biomarker. For example, a metric reflecting discrimination ability or accuracy is more relevant than one reflecting association for a candidate diagnostic biomarker. Detailed guidance for reporting a variety of types of health research studies is available on the EQUATOR website (http://www.equator-network.org/reporting-guidelines/). Several of particular relevance to biomarker studies are listed in Table 1. Although reporting guidelines do not dictate how research should be performed, many investigators find them helpful to consult when planning studies to be reminded of critical aspects of study design, conduct, and analysis to consider.
Table 1

Reporting guidelines particularly relevant to clinical biomarker research

AcronymReporting guideline titleWebsiteStudy type
BRISQBiospecimen reporting for improved study quality http://www.equator-network.org/reporting-guidelines/brisq/ Studies utilizing biospecimens
CONSORTConsolidated standards of reporting trials http://www.consort-statement.org/ Randomized clinical trials
REMARKReporting recommendations for tumor marker prognostic studies http://www.equator-network.org/reporting-guidelines/reporting-recommendations-for-tumour-marker-prognostic-studies-remark/ Tumor marker prognostic studies (and prognostic studies more generally)
STARDStandards for the reporting of diagnostic accuracy studies http://www.equator-network.org/reporting-guidelines/stard/ Diagnostic accuracy studies
STROBEStrengthening the reporting of observational studies in epidemiology http://www.equator-network.org/reporting-guidelines/strobe/ Observational studies in epidemiology (and more generally)
Reporting guidelines particularly relevant to clinical biomarker research

Results dissemination

The tendency to preferentially publish studies showing positive or statistically significant findings is known as publication bias. A related phenomenon is selective reporting of results within a study (e.g., only for certain outcome measures or subgroups among many examined), where usually those reported are statistically significant, especially in a desired or expected direction. Evidence for publication bias and selective reporting in clinical trials has been firmly established.9 For early clinical biomarker research, the potential for biases is greater due to lack of an organized system for study registration (analogous to ClinicalTrials.gov for clinical trials) and typical absence of comprehensive study protocols with prespecified statistical analysis plans. For every biomarker study reporting positive results, it is unknown how many studies of the same biomarker failing to achieve desired or statistically significant results never saw the light of day, or what resources were expended on failed or unreported studies. Although proposals have been made for biomarker study registration,10 resources to support registration systems are needed along with incentives or requirements from journals and funders, similar to existing mandates for registration of clinical trials in ClinicalTrials.gov.

CONCLUSION

A concerted effort involving many stakeholders is needed to provide guidance, resources, and incentives to successfully achieve research reproducibility goals. Signs of increased commitment to reproducibility are encouraging, but additional stakeholders will need to join the effort in order to succeed in changing the culture and improving reproducibility of early clinical biomarker research (see Supplementary Table S2). Table S1. Helpful references to guide assay analytical performance assessments Click here for additional data file. Table S2. Recent efforts aiming to enhance research reproducibility Click here for additional data file.
  9 in total

1.  Weighted analyses for cohort sampling designs.

Authors:  Robert J Gray
Journal:  Lifetime Data Anal       Date:  2008-08-19       Impact factor: 1.588

2.  Biases introduced by choosing controls to match risk factors of cases in biomarker research.

Authors:  Margaret Sullivan Pepe; Jing Fan; Christopher W Seymour; Christopher Li; Ying Huang; Ziding Feng
Journal:  Clin Chem       Date:  2012-06-22       Impact factor: 8.327

Review 3.  Biomarker studies: a call for a comprehensive biomarker study registry.

Authors:  Fabrice Andre; Lisa M McShane; Stefan Michiels; David F Ransohoff; Douglas G Altman; Jorge S Reis-Filho; Daniel F Hayes; Lajos Pusztai
Journal:  Nat Rev Clin Oncol       Date:  2011-03       Impact factor: 66.675

Review 4.  Tackling the widespread and critical impact of batch effects in high-throughput data.

Authors:  Jeffrey T Leek; Robert B Scharpf; Héctor Corrada Bravo; David Simcha; Benjamin Langmead; W Evan Johnson; Donald Geman; Keith Baggerly; Rafael A Irizarry
Journal:  Nat Rev Genet       Date:  2010-09-14       Impact factor: 53.242

5.  Clinical trials: discerning hype from substance.

Authors:  Thomas R Fleming
Journal:  Ann Intern Med       Date:  2010-09-21       Impact factor: 25.391

6.  A call for transparent reporting to optimize the predictive value of preclinical research.

Authors:  Story C Landis; Susan G Amara; Khusru Asadullah; Chris P Austin; Robi Blumenstein; Eileen W Bradley; Ronald G Crystal; Robert B Darnell; Robert J Ferrante; Howard Fillit; Robert Finkelstein; Marc Fisher; Howard E Gendelman; Robert M Golub; John L Goudreau; Robert A Gross; Amelie K Gubitz; Sharon E Hesterlee; David W Howells; John Huguenard; Katrina Kelner; Walter Koroshetz; Dimitri Krainc; Stanley E Lazic; Michael S Levine; Malcolm R Macleod; John M McCall; Richard T Moxley; Kalyani Narasimhan; Linda J Noble; Steve Perrin; John D Porter; Oswald Steward; Ellis Unger; Ursula Utz; Shai D Silberberg
Journal:  Nature       Date:  2012-10-11       Impact factor: 49.962

7.  Criteria for the use of omics-based predictors in clinical trials: explanation and elaboration.

Authors:  Lisa M McShane; Margaret M Cavenagh; Tracy G Lively; David A Eberhard; William L Bigbee; P Mickey Williams; Jill P Mesirov; Mei-Yin C Polley; Kelly Y Kim; James V Tricoli; Jeremy M G Taylor; Deborah J Shuman; Richard M Simon; James H Doroshow; Barbara A Conley
Journal:  BMC Med       Date:  2013-10-17       Impact factor: 11.150

Review 8.  Systematic review of the empirical evidence of study publication bias and outcome reporting bias.

Authors:  Kerry Dwan; Douglas G Altman; Juan A Arnaiz; Jill Bloom; An-Wen Chan; Eugenia Cronin; Evelyne Decullier; Philippa J Easterbrook; Erik Von Elm; Carrol Gamble; Davina Ghersi; John P A Ioannidis; John Simes; Paula R Williamson
Journal:  PLoS One       Date:  2008-08-28       Impact factor: 3.240

9.  Criteria for the use of omics-based predictors in clinical trials.

Authors:  Lisa M McShane; Margaret M Cavenagh; Tracy G Lively; David A Eberhard; William L Bigbee; P Mickey Williams; Jill P Mesirov; Mei-Yin C Polley; Kelly Y Kim; James V Tricoli; Jeremy M G Taylor; Deborah J Shuman; Richard M Simon; James H Doroshow; Barbara A Conley
Journal:  Nature       Date:  2013-10-17       Impact factor: 49.962

  9 in total
  10 in total

1.  SITC cancer immunotherapy resource document: a compass in the land of biomarker discovery.

Authors:  Siwen Hu-Lieskovan; Srabani Bhaumik; Kavita Dhodapkar; Jean-Charles J B Grivel; Sumati Gupta; Brent A Hanks; Sylvia Janetzki; Thomas O Kleen; Yoshinobu Koguchi; Amanda W Lund; Cristina Maccalli; Yolanda D Mahnke; Ruslan D Novosiadly; Senthamil R Selvan; Tasha Sims; Yingdong Zhao; Holden T Maecker
Journal:  J Immunother Cancer       Date:  2020-12       Impact factor: 13.751

Review 2.  Radiomics in immuno-oncology.

Authors:  Z Bodalal; I Wamelink; S Trebeschi; R G H Beets-Tan
Journal:  Immunooncol Technol       Date:  2021-04-16

Review 3.  Clinical trials in gynecologic oncology: Past, present, and future.

Authors:  Christina M Annunziata; Elise C Kohn
Journal:  Gynecol Oncol       Date:  2017-12-06       Impact factor: 5.482

Review 4.  Tutorial: best practices and considerations for mass-spectrometry-based protein biomarker discovery and validation.

Authors:  Ernesto S Nakayasu; Marina Gritsenko; Paul D Piehowski; Yuqian Gao; Daniel J Orton; Athena A Schepmoes; Thomas L Fillmore; Brigitte I Frohnert; Marian Rewers; Jeffrey P Krischer; Charles Ansong; Astrid M Suchy-Dicey; Carmella Evans-Molina; Wei-Jun Qian; Bobbie-Jo M Webb-Robertson; Thomas O Metz
Journal:  Nat Protoc       Date:  2021-07-09       Impact factor: 17.021

5.  Translating Precision.

Authors:  M A Pacanowski
Journal:  Clin Transl Sci       Date:  2017-02-13       Impact factor: 4.689

6.  An age-independent gene signature for monitoring acute rejection in kidney transplantation.

Authors:  Brian I Shaw; Daniel K Cheng; Chaitanya R Acharya; Robert B Ettenger; Herbert Kim Lyerly; Qing Cheng; Allan D Kirk; Eileen T Chambers
Journal:  Theranostics       Date:  2020-05-25       Impact factor: 11.556

7.  Considerations for Adapting Pre-existing Mechanistic Quantitative Systems Pharmacology Models for New Research Contexts.

Authors:  Michael Weis; Rebecca Baillie; Christina Friedrich
Journal:  Front Pharmacol       Date:  2019-04-18       Impact factor: 5.810

Review 8.  Peripheral and neural correlates of self-harm in children and adolescents: a scoping review.

Authors:  Victoria M Sparrow-Downes; Sara Trincao-Batra; Paula Cloutier; Amanda R Helleman; Mina Salamatmanesh; William Gardner; Anton Baksh; Rishi Kapur; Nicole Sheridan; Sinthuja Suntharalingam; Lisa Currie; Liam D Carrie; Arthur Hamilton; Kathleen Pajer
Journal:  BMC Psychiatry       Date:  2022-05-04       Impact factor: 3.630

Review 9.  Parkinson's Disease Biomarkers: Where Are We and Where Do We Go Next?

Authors:  Lana M Chahine; Matthew B Stern
Journal:  Mov Disord Clin Pract       Date:  2017-10-02

10.  Know Your Variability: Challenges in Mechanistic Modeling of Inflammatory Response in Inflammatory Bowel Disease (IBD).

Authors:  Katharine V Rogers; Indranil Bhattacharya; Steven W Martin; Satyaprakash Nayak
Journal:  Clin Transl Sci       Date:  2017-10-06       Impact factor: 4.689

  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.