Literature DB >> 24844265

Distinguishing between exploratory and confirmatory preclinical research will improve translation.

Jonathan Kimmelman1, Jeffrey S Mogil2, Ulrich Dirnagl3.   

Abstract

Preclinical researchers confront two overarching agendas related to drug development: selecting interventions amid a vast field of candidates, and producing rigorous evidence of clinical promise for a small number of interventions. We suggest that each challenge is best met by two different, complementary modes of investigation. In the first (exploratory investigation), researchers should aim at generating robust pathophysiological theories of disease. In the second (confirmatory investigation), researchers should aim at demonstrating strong and reproducible treatment effects in relevant animal models. Each mode entails different study designs, confronts different validity threats, and supports different kinds of inferences. Research policies should seek to disentangle the two modes and leverage their complementarity. In particular, policies should discourage the common use of exploratory studies to support confirmatory inferences, promote a greater volume of confirmatory investigation, and customize design and reporting guidelines for each mode.

Entities:  

Mesh:

Year:  2014        PMID: 24844265      PMCID: PMC4028181          DOI: 10.1371/journal.pbio.1001863

Source DB:  PubMed          Journal:  PLoS Biol        ISSN: 1544-9173            Impact factor:   8.029


Introduction

The past few years have witnessed growing consternation over the way researchers perform and report preclinical investigations of new drugs. The vast majority of drugs advanced into trials never recapitulate safety and efficacy observed in animal models, and these failures exact a heavy toll on trial volunteers, the research enterprise, and health care systems via higher drug prices. Because many preclinical studies poorly address internal validity threats [1], fail attempts at replication [2], are not published [3], or provide exaggerated estimates of clinical utility, numerous stakeholders are urging reforms in the way preclinical research is performed [4]. We would like to offer a cautionary perspective on these initiatives. We suggest that the ostensibly poor performance of many preclinical studies may in fact reflect strengths and intrinsic properties of what we call “exploratory investigation”—roughly, studies aimed at generating robust pathophysiological theories of disease. Policies aimed at improving translation should strive to preserve the extraordinary power of exploratory studies, which represent the majority of preclinical studies [5], while promoting a separate mode of clinical trial-like preclinical research, which we call “confirmatory” studies—that is, studies aimed at demonstrating strong and reproducible treatment effects in relevant animal models. We close by describing some ways of capitalizing on the complementarity of the two modes.

Exploratory Versus Confirmatory Research

Clinical translation of novel interventional strategies confronts two overarching challenges. First, researchers must negotiate a virtually unbounded landscape of potential targets, drugs, doses, and treatment regimens. A key task is to develop the theories, measurement techniques, and evidence for selecting a manageable number of interventions to carry forward. Second, clinical development is enormously expensive and exposes patients to unproven and possibly toxic interventions. Another key task of preclinical research is thus to produce evidence that is sufficiently compelling to warrant the economic and moral costs of clinical development. Overcoming these two challenges necessitates different modes of investigation. The first set of challenges is best met by studies that operate in the exploratory mode. We use “exploratory” to capture something broader than what is generally meant in statistics. In our conception, exploratory studies will aim primarily at developing pathophysiological theories that enable pursuit of different approaches. Exploratory studies tend to consist of a package of small and flexible experiments using different methodologies, including molecular and cellular analyses. These individual experiments may or may not employ inferential statistics. Exploratory studies are often driven by a series of hypotheses that are either loosely articulated or that evolve over the course of sequential experiments. Often, exploratory studies include tests of an intervention's efficacy against disease in live animals as a way of validating the pathophysiological theories (“efficacy studies”). Neither the sequence of individual experiments in exploratory studies, nor details of their design (including sample size, since effect sizes may be unknown), is necessarily established at the outset of investigation. The second set of challenges is best overcome by studies that operate in a confirmatory mode. Such studies will resemble adequately powered clinical trials, and consist mainly of “efficacy studies” that use rigid and pre-specified designs, a priori stated hypotheses, prolonged durations, and the most clinically relevant assays and endpoints available. These studies aim less at elaborating theories or mechanisms of a drug's action than rigorously testing a drug's clinical potential and restricting the advance of ineffective interventions advanced into clinical testing. Exploratory studies are a complement to confirmatory studies in that the former generates precisely articulated hypotheses about drug effects that can be put to “crucial testing” in the latter before clinical development. Currently, the vast majority of preclinical studies more closely resemble exploratory studies, although a small but growing number of studies operate in a confirmatory mode. These different orientations carry important imperatives for the design, reporting, error tendencies, and application of preclinical studies. What may be an inferential strength for exploratory study can be a hindrance or even a fatal flaw for confirmatory studies and vice versa. Policies and practices aimed at improving clinical translation should recognize at least four major contrasts between the two modes of investigation.

Implications for Design and Valid Interpretation

The first difference has already been noted: whereas exploratory studies should mainly aim at deriving or testing theoretical claims, confirmatory studies should test clinical utility of new interventions. Since theories are not directly observable, they are tested by assembling corroboratory evidence across different lines of experimentation. This theoretical orientation in preclinical research is reflected in the fact that a good part of the acreage in publications is devoted to molecular or cellular analyses (e.g., gene expression, immunohistochemistry, electrophysiology), not efficacy studies. Spreading proof across different lines of experiment—a process called “conceptual replication” [6]—has several consequences for predictive value. On the one hand, threats to the validity of theoretical claims driving a preclinical study are mitigated—though not eliminated—by conceptual replications. On the other hand, therapeutic claims arising from efficacy studies contained in the exploratory package will be prone to larger random and systematic variation: such studies invest less in any single experiment, and therefore employ smaller sample sizes and less fastidious designs. In contrast, because confirmatory studies “bet the house” on a single, pivotal efficacy study and measurement technique, there is more at stake scientifically in minimizing random and systematic error. Second, whereas exploratory studies should place a premium on sensitivity (i.e., detecting all strategies that might be useful), confirmatory studies should be more concerned with specificity (i.e., excluding all strategies that will prove useless in clinical trials). This is because the task of exploration is to catch a small number of promising theories, targets, compounds, doses, or variants of a target indication against a large field. However, in many areas of drug development, the prior probability of discovering useful strategies is extraordinarily low. This means that even in the ideal, where exploratory studies have very high sensitivity and specificity, most candidates that are declared promising will represent false positives. Since there are large financial and human costs for advancing these false positives into trials, the task of confirmation is to eliminate “false positives” that are captured in exploration. Further, the agonizingly low positive predictive value of exploratory studies may have as much to do with base rates as it does with bias. Third, use of small sample sizes for efficacy experiments contained in exploratory studies may lead to large random variation that produces the appearance of bias even in its absence. This dynamic, known as the “winner's curse” [7], reflects the fact that research in the exploratory mode will often test many different strategies in parallel, and this is only feasible if small sample sizes are used. As a consequence of random variation alone, some experiments will produce larger effects that regress to the mean if replicated. In contrast, confirmatory studies should employ sufficiently large sample sizes as to minimize the effect of random variation, such that dwindling effect sizes on replication may be symptomatic of publication bias rather than natural regression. Last, exploratory studies often involve testing interventions alongside techniques used to measure their effects. In contrast, methods should be well established when an intervention is tested in confirmatory studies. Assays for testing pathophysiological responses, or the probative value of biomarkers, or skills for performing a behavioral test may be still in development at the point of exploratory investigation. One example of this is uncertainty surrounding techniques for testing drugs that target cancer stem cells. Here, standard assays for testing the clinical promise of cancer drugs are almost useless, yet there is little consensus about which assays to use instead [8]. Another example might be where a graduate student conducts experiments before having mastered the requisite manual skills. As a consequence of uncertainty surrounding measurement, exploratory researchers encounter difficulty discriminating informative and uninformative findings: “positive” findings may be attributable to assay artefacts; “negative” findings may reflect defects in the measurement tools, choice of the wrong treatment regimen, or suboptimal experimenter skill. Since the value of uninformative findings for the broader research community is limited, the absence of firm rules for discrimination legitimately confounds decisions about what findings to publish and how to interpret them. Any blanket proscription against “hiding” data risks obscuring truly interesting findings amid a large volume of studies that the experimenter knows to be uninformative to the broader research community: “practice runs,” experiments on miscalibrated instruments, or findings using methods that are later discovered to be error prone. On the other hand, where researchers have grounds for confidence in the regimens for testing, nonpublication of negative findings represents a demonstrable breach of scientific integrity. This will tend to be a much greater concern in confirmatory testing, since measurement techniques tend to be more established in that setting. In sum, there are many factors that explain why preclinical studies are prone to producing “false positives” or outcome patterns that give the appearance of bias. Yet to some degree, these reflect strengths of exploratory research, such as its ability to narrow the field of intervention candidates using an economy of resources, to select among myriad pathophysiological theories, and to hone techniques of measuring clinical promise. These are necessary precursors to the sorts of rigorous confirmatory experiments that should be used to justify clinical development.

Improving Design and Interpretation of Preclinical Research

Though some of the above contrasts may appear obvious to anyone with a basic understanding of statistics and experimental design, they are not adequately reflected in many reforms urged by critics of preclinical research—e.g., calls for using larger sample sizes, “gold standard” animal models, or independent replication [9],[10]. Some proposals entail non-trivial burdens such as restructuring laboratory practices, writing up and/or depositing inconclusive findings, or using larger sample sizes, and hence undermine the economy of exploratory activities. Reforms are more likely to have a transformative impact on drug development if researchers can capitalize on the complementary properties of both exploratory and translational studies, and tailor study design, reporting, and application of findings accordingly. To that end, we offer three sets of recommendations. First, all protocols and publications should pre-specify whether they are “exploratory” or “confirmatory” studies, with the latter category reserved for studies that aim at demonstrating promise of clinical utility for an intervention. We note that other commentators have made similar calls [11],[12]. Journal editors and funding agencies should promote this demarcation by requiring it for submitted manuscripts and grants. Standards for review should then hinge on the way investigators classify studies. For instance, confirmatory studies should be held to internal and construct validity standards similar to those used in clinical trials: studies should address confounders like sample or observation bias, use pre-specified statistical analyses, match the experimental design to the conditions where findings are expected to be applied, and report findings in ways that enable meaningful interpretation by non-experts. Large sample sizes, fastidious experimental conditions, and conservative statistical analyses may be counterproductive for exploration. Instead, exploratory studies should be evaluated on the basis of whether findings using disparate and methodologically sound lines of investigation are coherent and fecund. Second, the research community should devise mechanisms for coupling confirmatory studies to exploratory ones. As noted above, only a small minority of preclinical studies are put to confirmatory testing. Once intervention strategies are discovered in exploration, those wishing to launch clinical development should be expected to run, or at least reference, stand-alone confirmatory studies before launching trials [10],[13]. One way of promoting this would be for oversight bodies—Research Ethics Boards, public funding agencies, and regulators—to condition approval of any trial delivering putatively active drug doses on positive preclinical confirmatory studies. Like clinical trials, such studies should prospectively register, adhere to (and preferably publish) protocols, and report findings according to standards and regardless of effect sizes. As human trial findings are much more informative when they are embedded within a web of related findings [14], medical journals should require that investigators deposit confirmatory preclinical findings when they accept for publication trials involving efficacy primary endpoints. Third, many recommendations and mechanisms for improving preclinical study design are mainly suited for confirmatory studies. Some recommendations—like calls for more regular replication or simple measures to reduce factors like observer bias (e.g., randomization and assessor blinding) —are sensible across both modes of investigation. Others seem more suited for confirmatory studies and may be counterproductive for exploratory studies. Use of larger sample sizes and prospective registration, for example, involve additional investments, infrastructure, and compliance burdens. The former means sacrifice of more animals than necessary to identify promising strategies. The latter would be very taxing for researchers, since public disclosures early on in a research program would invite free-riding; moreover, registration of exploratory studies offers little to a research community if researchers themselves have significant doubts about measurement techniques. Perhaps the largest validity threats in exploratory research reside not in efficacy studies, but in the withholding of findings that disrupt the coherence of theoretical claims, in the assembly of theories that build on a series of falsely positive experimental results, or in the nonperformance of replication experiments because of insufficient incentive. The research community has much to gain from guidelines and mechanisms that specifically address such tendencies in exploration. One place to start would be to establish data ontologies and databases for deposition of exploratory findings so that discordant findings can be accessed. Another would be the creation of mechanisms that encourage confirmatory studies. For example, several journals now solicit bids for replication studies from the research community, guaranteeing the winning bidder publication on successful completion of the study. Others maintain “results blind” publication categories, where reviews are based on a submitted protocol rather than effect sizes [15],[16]. Journals might also encourage researchers to deposit replications or experiments that are discordant with published exploratory studies, but that are insufficient to constitute a new paper, by creating a section for very short research reports that consist of a single experiment or an attempted replication. According to influential accounts of the research process, science flourishes best when researchers pursue different agendas, harboring different biases [17]. Preclinical research, in particular, entails two complimentary agendas: one is to narrow a large field of potential therapies by refining pathophysiological theories of disease, and the other is to generate reliable evidence of a therapy's clinical utility in a proxy species. Each encounters different constraints and validity threats. The key to improving preclinical research is devising practices that leverage one to the advantage of the other.
  14 in total

1.  Replication and reproducibility in spinal cord injury research.

Authors:  Oswald Steward; Phillip G Popovich; W Dalton Dietrich; Naomi Kleitman
Journal:  Exp Neurol       Date:  2011-11-10       Impact factor: 5.330

2.  Should preclinical studies be registered?

Authors:  Jonathan Kimmelman; James A Anderson
Journal:  Nat Biotechnol       Date:  2012-06-07       Impact factor: 54.908

3.  An Agenda for Purely Confirmatory Research.

Authors:  Eric-Jan Wagenmakers; Ruud Wetzels; Denny Borsboom; Han L J van der Maas; Rogier A Kievit
Journal:  Perspect Psychol Sci       Date:  2012-11

Review 4.  Power failure: why small sample size undermines the reliability of neuroscience.

Authors:  Katherine S Button; John P A Ioannidis; Claire Mokrysz; Brian A Nosek; Jonathan Flint; Emma S J Robinson; Marcus R Munafò
Journal:  Nat Rev Neurosci       Date:  2013-04-10       Impact factor: 34.870

5.  A call for transparent reporting to optimize the predictive value of preclinical research.

Authors:  Story C Landis; Susan G Amara; Khusru Asadullah; Chris P Austin; Robi Blumenstein; Eileen W Bradley; Ronald G Crystal; Robert B Darnell; Robert J Ferrante; Howard Fillit; Robert Finkelstein; Marc Fisher; Howard E Gendelman; Robert M Golub; John L Goudreau; Robert A Gross; Amelie K Gubitz; Sharon E Hesterlee; David W Howells; John Huguenard; Katrina Kelner; Walter Koroshetz; Dimitri Krainc; Stanley E Lazic; Michael S Levine; Malcolm R Macleod; John M McCall; Richard T Moxley; Kalyani Narasimhan; Linda J Noble; Steve Perrin; John D Porter; Oswald Steward; Ellis Unger; Ursula Utz; Shai D Silberberg
Journal:  Nature       Date:  2012-10-11       Impact factor: 49.962

6.  Publication bias in reports of animal stroke studies leads to major overstatement of efficacy.

Authors:  Emily S Sena; H Bart van der Worp; Philip M W Bath; David W Howells; Malcolm R Macleod
Journal:  PLoS Biol       Date:  2010-03-30       Impact factor: 8.029

7.  Heterogeneity in cancer: cancer stem cells versus clonal evolution.

Authors:  Mark Shackleton; Elsa Quintana; Eric R Fearon; Sean J Morrison
Journal:  Cell       Date:  2009-09-04       Impact factor: 41.582

8.  Pain research from 1975 to 2007: a categorical and bibliometric meta-trend analysis of every Research Paper published in the journal, Pain.

Authors:  Jeffrey S Mogil; Kent Simmonds; Maureen J Simmonds
Journal:  Pain       Date:  2008-12-27       Impact factor: 6.961

9.  A concerted appeal for international cooperation in preclinical stroke research.

Authors:  Ulrich Dirnagl; Antoine Hakim; Malcolm Macleod; Marc Fisher; David Howells; Stuart M Alan; Gary Steinberg; Anna Planas; Johannes Boltze; Sean Savitz; Costantino Iadecola; Stephen Meairs
Journal:  Stroke       Date:  2013-04-18       Impact factor: 7.914

Review 10.  Threats to validity in the design and conduct of preclinical efficacy studies: a systematic review of guidelines for in vivo animal experiments.

Authors:  Valerie C Henderson; Jonathan Kimmelman; Dean Fergusson; Jeremy M Grimshaw; Dan G Hackam
Journal:  PLoS Med       Date:  2013-07-23       Impact factor: 11.069

View more
  65 in total

1.  The use of systematic reviews and reporting guidelines to advance the implementation of the 3Rs.

Authors:  Marc T Avey; Nicole Fenwick; Gilly Griffin
Journal:  J Am Assoc Lab Anim Sci       Date:  2015-03       Impact factor: 1.232

2.  The p value wars (again).

Authors:  Ulrich Dirnagl
Journal:  Eur J Nucl Med Mol Imaging       Date:  2019-11       Impact factor: 9.236

3.  The Interplay of Ethics, Animal Welfare, and IACUC Oversight on the Reproducibility of Animal Studies.

Authors:  Stacy L Pritt; Robert E Hammer
Journal:  Comp Med       Date:  2017-03-01       Impact factor: 0.982

4.  Interaction of ARC and Daxx: A Novel Endogenous Target to Preserve Motor Function and Cell Loss after Focal Brain Ischemia in Mice.

Authors:  Stefan Donath; Junfeng An; Sabrina Lin Lin Lee; Karen Gertz; Anna Lena Datwyler; Ulrike Harms; Susanne Müller; Tracy Deanne Farr; Martina Füchtemeier; Gisela Lättig-Tünnemann; Janet Lips; Marco Foddis; Larissa Mosch; René Bernard; Ulrike Grittner; Mustafa Balkaya; Golo Kronenberg; Ulrich Dirnagl; Matthias Endres; Christoph Harms
Journal:  J Neurosci       Date:  2016-08-03       Impact factor: 6.167

5.  Reproducibility of animal research in light of biological variation.

Authors:  Bernhard Voelkl; Naomi S Altman; Anders Forsman; Wolfgang Forstmeier; Jessica Gurevitch; Ivana Jaric; Natasha A Karp; Martien J Kas; Holger Schielzeth; Tom Van de Casteele; Hanno Würbel
Journal:  Nat Rev Neurosci       Date:  2020-06-02       Impact factor: 34.870

6.  Accelerating Biomedical Discoveries through Rigor and Transparency.

Authors:  Judith A Hewitt; Liliana L Brown; Stephanie J Murphy; Franziska Grieder; Shai D Silberberg
Journal:  ILAR J       Date:  2017-07-01

7.  Clinically approved IVIg delivered to the hippocampus with focused ultrasound promotes neurogenesis in a model of Alzheimer's disease.

Authors:  Sonam Dubey; Stefan Heinen; Slavica Krantic; JoAnne McLaurin; Donald R Branch; Kullervo Hynynen; Isabelle Aubert
Journal:  Proc Natl Acad Sci U S A       Date:  2020-12-07       Impact factor: 11.205

8.  The case for introducing pre-registered confirmatory pharmacological pre-clinical studies.

Authors:  Olivia Kiwanuka; Bo-Michael Bellander; Anders Hånell
Journal:  J Cereb Blood Flow Metab       Date:  2018-02-26       Impact factor: 6.200

9.  Moving Sport and Exercise Science Forward: A Call for the Adoption of More Transparent Research Practices.

Authors:  Aaron R Caldwell; Andrew D Vigotsky; Matthew S Tenan; Rémi Radel; David T Mellor; Andreas Kreutzer; Ian M Lahart; John P Mills; Matthieu P Boisgontier
Journal:  Sports Med       Date:  2020-03       Impact factor: 11.136

10.  T-Cell Mediation of Pregnancy Analgesia Affecting Chronic Pain in Mice.

Authors:  Sarah F Rosen; Boram Ham; Shannon Drouin; Nadia Boachie; Anne-Julie Chabot-Dore; Jean-Sebastien Austin; Luda Diatchenko; Jeffrey S Mogil
Journal:  J Neurosci       Date:  2017-09-06       Impact factor: 6.167

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.