Literature DB >> 35854983

The Value of the Information That Can Be Generated: Optimizing Study Design to Enable the Study of Treatments Addressing an Unmet Need for Rare Pathogens.

Aaron Dane1, John H Rex2, Paul Newell2, Nigel Stallard3.   

Abstract

In traditional phase 3 trials confirming safety and efficacy of new treatments relative to a comparator, a 1-sided type I error rate of 2.5% is traditionally used and typically leads to minimum sizes of 300-600 subjects per study. However, for rare pathogens, it may be necessary to work with data from as few as 50-100 subjects. For areas with a high unmet need, there is a balance between traditional type I error and power and enabling feasible studies. In such cases, an alternative 1-sided alpha level of 5% or 10% should be considered, and we review herein the implications of such approaches. Resolving this question requires engagement of patients, the medical community, regulatory agencies, and trial sponsors.
© The Author(s) 2022. Published by Oxford University Press on behalf of the Infectious Diseases Society of America.

Entities:  

Keywords:  alternate statistical criteria; antibiotics; antifungals; small trials; unmet need

Year:  2022        PMID: 35854983      PMCID: PMC9290570          DOI: 10.1093/ofid/ofac266

Source DB:  PubMed          Journal:  Open Forum Infect Dis        ISSN: 2328-8957            Impact factor:   4.423


Traditional phase 3 trials confirm safety and efficacy of a new treatment relative to an active comparator or placebo using a 1-sided type I error rate of 2.5%. Although the 2.5% type I error rate is not a regulatory requirement, it is traditionally used by regulatory agencies to limit false conclusions of treatment effect. Given the high unmet need, in some infectious disease research the US Food and Drug Administration (FDA) have considered flexibility in benefit-risk considerations, granting approval with a single adequate and well controlled trial, with supportive information such as a phase 2 trial or information from in vitro studies and animal models of infection. Although this enables progression of some clinical development programs, there are still situations when it may only be feasible to recruit 50–100 patients in many years [1], for example, the plazomicin bacteremia study for carbapenem-resistant Enterobacteriaceae [2], making even a single trial extremely challenging. As a result, there are no, or very limited, data in such patient populations, so development of agents in small populations requires alternative approaches. More importantly, we only envision these approaches being used when a larger trial is simply not possible, so these ideas do not apply when there is a realistic choice between a larger or smaller study. Drawing on value of information ideas within the orphan drug area [3-6], we propose an approach that uses alternative type I error rates in a decision-making and sample-size determination framework in which typically sized controlled studies are extremely challenging to enroll. The goal is to balance the risks of correct and incorrect decisions from limited data. Both approval and nonapproval carry risks, and we propose an approach that considers risks of failing to approve a new effective medicine, approval of an ineffective medicine, and conducting a large randomized controlled trial ([RCT] with some patients receiving suboptimal therapy) in a limited population. We discuss regulatory approval on the basis of efficacy, because the aim is to understand the risks of using alternative statistical criteria within an efficacy assessment. The safety and benefit-risk profile are also critically important and need to be considered in any evaluation, but because the safety profile is unique to each treatment, we have not discussed this aspect in detail, but it is always a critical aspect of any evaluation of the evidence to support regulatory approval.

Superiority Trials Are Not Appropriate for All Development Programs

It is often perceived that demonstration of superiority in small trials with resistant pathogens should be straightforward if a strong effect exists. However, this is infeasible because we cannot study ineffective comparators in seriously ill patients, and demonstration of superiority is challenging when therapies exist with some efficacy, albeit suboptimal [7, 8]. Although combinations of less effective therapies may have toxicities, their existence makes superiority hard to demonstrate. In some settings, the use of a nonstandard significance level to assess superiority may make studies more feasible, but in many cases the issues associated with demonstrating superiority described above will still exist, meaning alternative options are also required. Therefore, we focus on noninferiority (NI) trials, but similar arguments to those presented could be used to assess the impact of changes to type I error in superiority trials.

Decision Making and Sample Size in Randomized Controlled Trials With Rare Pathogens

The challenge reduces to recognizing that rare pathogens make it impossible for practical purposes to conduct either a small superiority study or a large noninferiority active-control study. However, new therapies are needed for rare pathogens, and we can approach this problem by being sure that we are confident of the following: (1) we will approve effective treatments (showing superiority or noninferiority with an active control); and (2) we will not approve ineffective treatments. In registrational studies, it is traditional to set power (probability of correctly showing the new agent is effective) at 80%–90% and type I error at 2.5% (1-sided, the probability of concluding a new treatment is effective when it is not) and set sample size accordingly [9]. Table 1 summarizes these concepts and the consequences for patients within the trial and those beyond the trial who will receive the current or new therapy dependent upon the trial results.
Table 1.

Scenarios to Consider When Analyzing Decision Making With Small Datasets

True Response Rate or Mortality RateConsequences Within RCTConsequences for Patients After RCT
New agent is worse than controlEvery patient randomized to new agent risks a worse outcomeIf the new agent is approved, this problem is perpetuated, and all future patients would receive a less effective therapyWe should avoid incorrect approvals so want to have low power of approval
New agent is better than controlEvery patient randomized to control risks a worse outcomeIf the new agent is not approved, this problem is perpetuated, and all future patients would receive a less effective therapyWe should retain high power of approval
New agent is similar to controlEvery patient has approximately the same outcomeEither existing treatment or a new therapy would be acceptable, but we still need to have additional therapies available. “Therefore, we still want to retain high power of approval

Abbreviations: RCT, randomized controlled trial.

Scenarios to Consider When Analyzing Decision Making With Small Datasets Abbreviations: RCT, randomized controlled trial. When we design the trial, we do not know the true response rate, so we must understand the probability of approval for different study sizes and scenarios regarding true treatment effects. We are attempting to identify a “sweet spot” for sample size balancing the type I error and power when it is infeasible to obtain the traditional chances of correctly or incorrectly approving a new treatment. Where possible, trials with conventional power and type I error are preferable. However, if the trial is too large, this deprives patients of timely access to an effective therapy and means that while the trial is ongoing, some patients within the trial will receive an inferior treatment. If a trial is too large, developing the therapy is infeasible because we cannot generate data for that patient population. In contrast, if a trial is too small, it increases the probability of making an incorrect decision (either approval or nonapproval). We propose that when large trials are infeasible, alternative type I error rates should be explicitly considered within the context of the benefits and risks of new treatments, particularly when accompanied by supportive information such as data from in vitro studies and animal models of infection. We consider this for rare pathogens in instances in which new therapies are needed and patient numbers make clinical trials prohibitively lengthy or simply impossible to undertake. This has the potential to accelerate availability of new effective treatments rather than condemn patients to suboptimal treatment while larger trials are either attempted in vain or never attempted.

Understanding Clinical Trial Risks

In considering alternative statistical criteria, we first state key risks and then consider acceptable risk levels in an area of extreme unmet need. Risks must be considered for a range of scenarios in which the true effect of the test agent related to control is unknown. This is true of all studies but may need to be considered differently when trials cannot be designed to standard levels of power and type I error. Important scenarios are situations in which the new agent is better than, worse than, or similar to control. We consider these in turn for a randomized clinical trial using a 1:1 randomization to new agent or control with a 20% NI margin and 95% confidence interval [CI] (2-sided alpha = 0.05), as has been used for some trials [10, 11].

Scenario 1: New Agent Better Than Control

Imagine the new agent and control have true response rates (unknown before the trial is conducted) of 60% and 40%, respectively. The correct outcome would be to conclude NI and approve the new agent. The probability that NI is concluded (power) with this scenario increases with sample size (N) (Figure 1A) but does not improve greatly with more than 40–50 patients/group, when power approaches 90%. Although more data may be needed from a safety perspective, this would be sufficient to show the new agent has efficacy noninferior to control.
Figure 1.

Plot of power when using 95% confidence interval and −20% noninferiority margin. Cont, control.

Plot of power when using 95% confidence interval and −20% noninferiority margin. Cont, control.

Scenario 2: New Agent Slightly Better Than Control

When the new agent and control have true response rates of 50% and 40%, respectively, the correct outcome would again be to conclude NI and approve the new agent. In this case, as presented in Figure 1B, the probability NI is concluded increases with sample size, but power is lower for a given sample size than for Scenario 1 and approaches 80% at 50 patients/group.

Scenario 3: New Agent Similar to Control

In the third scenario, the true (but as yet unknown) response rates are the same (40%) for new agent and control. The correct decision in this scenario is more nuanced, and we must consider the value of currently available therapies, even if these are suboptimal. As stated, we assume that it would be desirable to have additional therapies available. In this case, the probability of concluding NI (power) increases with sample size (Figure 1C) but requires a larger sample size. A study with ∼50 patients/group has ∼50% power, whereas a study of 80 patients/group provides ∼80% power. It would take a large study to achieve 90% power. Because it is unlikely that studies would be undertaken with only a 50/50 chance of success, developing new agents is not possible when it is infeasible to recruit more than 100 patients. If new therapies are badly needed, a discussion on required levels of type I error and power should be considered.

Scenario 4: New Agent Slightly Worse Than Control

When the true response rates for the new agent and control are 30% and 40%, respectively, the correct outcome and associated approval is less obvious. If the primary endpoint of the study relates to all-cause mortality, you would not want to approve such an agent, but with a response rate endpoint, if there is a high unmet need, the available treatment has significant toxicity, and there is increasing resistance to this available treatment there could still be value in a new agent. You would not always want to approve such an agent, but consideration of these components of the available therapy would be important when considering the need for new agents. Therefore, with this scenario, there is no correct outcome for every situation. Rather, consideration of whether the new agent should or should not be approved will depend upon the reason why a new agent is needed. In this case, the plot shows the probability of concluding NI is <25% for sizes of < ∼75/arm (Figure 1D). The probability of concluding NI increases as the sample size further increases, and at 200/arm the probability approaches 50%. However, as such sample sizes are presumed impossible for rare pathogens this again shows a relatively small chance of concluding NI with achievable sample sizes of 40–50 patients/group.

Scenario 5: New Agent Worse Than Control

When the true (but as yet unknown) response rates for the new agent and control are 20% and 40%, respectively, we assume we would definitely not want to approve the new agent, so the correct outcome is that the new agent should not be approved. In this case, the probability of concluding NI is always very low (Figure 1E). Because the true difference between treatments equals the NI margin, this is the type I error rate (2.5%), so we rarely conclude NI erroneously, whatever the study size. Therefore, 40–50 patients/group again seems reasonable.

The Effect on Future Patients With This Rare Condition

We have shown that trials studying infections caused by rare pathogens are infeasible with traditional power and type I error rates. Another important component is the implication of any drug approval decision on the treatments received by patients with this rare pathogen, both in the RCT and in the future. As a hypothetical example, suppose we have an overall population of 1000 patients who might benefit from a new treatment, and from this population we recruit 100 patients (50/arm) to a trial comparing the new agent with control, the results of which will define the therapy used for the remaining 900 patients, dependent upon the NI conclusion made from the trial (see Table 2).
Table 2.

Expected Number of Responses in the Entire Population With an NI Trial of 100 or 400 Patients: Response Rates of 60% for the New Agent vs 40% for Control

RCT Sample SizeResponses in RCTProbability New Agent Is ApprovedResponses After Trial
New AgentControlIf New Agent Is ApprovedIf New Agent Not Approved (Control Is Used)
50/arm30/5020/500.983540/900360/900
Expected Successes in Overall Population:30 + 20 + (0.983 × 540) + (0.017 × 360) = 587 (58.7% of all patients)
200/arm120/20080/2000.999360/600240/600
Expected Successes in Overall Population:120 + 80 + (0.999 × 360) + (0.001 × 240) = 560 (ie, 56% of all patients)

Abbreviations: NI, noninferiority; RCT, randomized controlled trial.

Expected Number of Responses in the Entire Population With an NI Trial of 100 or 400 Patients: Response Rates of 60% for the New Agent vs 40% for Control Abbreviations: NI, noninferiority; RCT, randomized controlled trial. Assuming all patients are included in the trial or are treated based on the results of the trial, we can calculate how many patients are successfully treated using the following. (1) How many of 50 trial patients on the new agent are successfully treated? (2) How many of 50 trial patients on Control are successfully treated? (3) How many of remaining 900 patients are successfully treated depending upon the trial outcome? The expected number of successes can be calculated under the scenarios considered above. When the new agent response rate is 60%, we would expect 30 of 50 trial patients receiving the new agent and 20 of 50 trial patients receiving control to respond (Table 2). We would expect 540 of the remaining patients to respond if NI is concluded and the new agent is used and 360 of the 900 patients to respond on control if NI were not concluded. Using a 20% NI margin and 95% CI, with true response rates of 60% and 40%, a trial with 50 patients/arm would have a 98.3% chance of concluding NI. Thus, we expect the new agent to be approved 98.3% of the time (so 540 of 900 future patients would respond) and the control to be used (so 360 of 900 respond) 1.7% of the time. The total expected number of responses from the 1000 patients is therefore 587 as displayed in the first row of Table 2. The second row of Table 2 shows similar calculations using the same assumptions but a larger RCT including 200 patients/group, leaving 600 patients treated based on the RCT results. In this case, the total expected number of responses is lower at 560 patients. This may seem counterintuitive, but although larger RCTs increase the probability of reaching the correct conclusion, they also mean we wait longer to reach this conclusion, meaning more patients within the trial are randomized to receive a less effective therapy, and a smaller number of patients benefit from the more effective therapy after the trial. The example in Table 2 is for the most optimistic scenario considered and presents results for trials with 50 and 200 patients per arm only. Similar calculations are possible for each of the 5 scenarios described and for different sample sizes and show a similar pattern of declining response with larger sample sizes for less optimistic scenarios (Figure 2). Recall that Figure 1 shows the probability of concluding NI in each case, whereas Figure 2 shows the proportion of responses expected in the population for each scenario. These results can be used to assess whether the risks of correctly approving an effective therapy, incorrectly approving an ineffective therapy, and the number of patients receiving suboptimal treatment have been appropriately balanced for a given trial size.
Figure 2.

Plot of expected (Exp.) proportion (prop.) of responses in a trial using 95% confidence interval and −20% noninferiority margin. Cont, control.

Plot of expected (Exp.) proportion (prop.) of responses in a trial using 95% confidence interval and −20% noninferiority margin. Cont, control. If the new agent is better than the control, a trial with 40 patients/group gives reasonable power and optimizes the expected successes in the overall population based on the treatment patients receive during and after the trial (Figure 2A). When the new agent is slightly better than control, the power is lower than when the new agent is clearly better, but it still declines after 40–50 patients/arm. If the new agent is worse than control, as noted above, the probability of incorrectly recommending the new agent is low irrespective of the RCT sample size. The expected proportion of responses decreases with increasing sample size because more patients in the trial receive the (less effective) new agent (Figure 2E). When the new agent is slightly worse than control, a similar pattern is seen with the expected proportion of successfully treated patients, whereas the probability of concluding NI increases with increasing sample size, but there is still only 15%–20% probability of concluding NI with 40–50 patients/arm. If the new agent and control have similar efficacy, the probability of recommending the new agent increases with N, and a larger study is required using traditional criteria. In this scenario, because the new agent and control have the same response rate, the expected proportion of patients to respond is constant (here 40%) (Figure 2C). Although the unmet need may appear less urgent when the efficacy of the new agent is similar to control, there are still a number of cases in which additional treatment options are urgently needed due to emerging resistance or toxicity concerns with current therapies. In this scenario, a study including ∼100 patients/group would be required to demonstrate efficacy. In studies of some rare or resistant pathogens, even 40–50 patients/group may be challenging, so it is necessary to find a way of running a feasible trial. We next suggest a framework for assessing alternative statistical criteria building on ideas presented above.

A Framework for Using Alternative Statistical Criteria

We have shown that although conventional error rates are achieved using 40–50 patients/arm when the new treatment is truly clearly better or worse than the control, 100 patients/arm may be necessary when the 2 treatments have similar efficacy. When such a study is infeasible, and additional treatment options are desirable, use of an alternative type I error rate (or CI) can be considered. Here we present the use of an 80% CI (2-sided alpha = 0.2) and a 20% NI margin. This approach would be reserved for circumstances with high unmet need and limited population where it is critical that new therapies are available but could not otherwise be developed. We will not discuss how an agent would meet these criteria, but we do outline how the risks associated with failure to make new therapies available, and the risk of incorrectly approving new agents, could be considered for such agents. Properties of trials with an 80% CI are shown in Figure 3, which includes in one plot the type of information presented in Figures 1 and 2. When the new agent is better than control, the expected number of patients responding declines after ∼50 patients/arm. The probability of a successful trial when the treatments are similar is higher (Figure 3, Panel C), approaching 80% with 50–60 patients/arm. The probability of such a trial incorrectly concluding NI when the new agent is worse (Figure 3, Panel E) is 10%, so it is above the usual value of 2.5% when using a 95% CI. This should be a point for discussion among patients, Infectious Diseases physicians, regulatory agencies, and sponsors for rare pathogens or for limited populations when additional treatment options are desperately needed. When the new agent is slightly worse, a similar pattern is seen to that when using a 95% CI in that the probability of concluding NI increases with increasing sample size, in this case meaning the chance of concluding NI approaches 40% with 50 patients/arm. As stated previously, the utility of an agent with the possibility of a slightly lower response rate but the ability to provide other benefits in terms of less toxicity or pathogen resistance should form part of the assessment of using an 80% CI.
Figure 3.

Plot of power and expected (Exp.) proportion (prop.) of responses in a trial using 80% confidence interval and −20% noninferiority margin. Cont, control.

Plot of power and expected (Exp.) proportion (prop.) of responses in a trial using 80% confidence interval and −20% noninferiority margin. Cont, control. The previous sections showed that when the new agent is clearly superior to the control, this would be apparent from 50 patients/arm. The alternative type I error rate level is important when the new agent and control are similar or slightly inferior. If it is only feasible to run a study of 100 patients, a trial with traditional type I error rate is impossible. Additional flexibility is then required, particularly when available therapies are suboptimal, such as being the drug of last resort or with toxicity issues where it is desirable to have additional therapies available. In this case, it will often be a choice between a smaller dataset or having no data at all.

DISCUSSION

Although studies of rare pathogens are challenging, the development of agents for such pathogens remains important. We have shown that although studies with traditional statistical criteria may be impossible, we can articulate a decision-making framework to support smaller studies. The proposed framework displays the risks of incorrect approval of an ineffective agent and incorrect rejection of an effective agent based on efficacy, along with the expected number of successes within a small population, highlighting that agreement needs to be reached across the community on how to balance these factors. The approach is similar to that proposed for orphan drug trials [3-6] in which a type I error rate larger than 0.025 is sometimes used [12, 13]. We have shown that comparative trials with 50 patients/arm can provide reliable decisions and could be the key to enabling any development for patients with rare pathogens. In our example, we have assumed that the control group event rate is 40%, and although our exact findings regarding the probability of approval may not hold if the control group rate were very different, similar methods could be used to choose an appropriate sample size. We have also shown that a larger study does not always provide better outcomes in a limited population because fewer patients remain to benefit from a therapy after the trial. In addition, in a trial with a 1:1 randomization in which one treatment is less effective, 50% of patients receive this suboptimal therapy. We have presented an example in which an 80% CI is used alongside a 20% NI margin. Dependent upon the degree of unmet need, feasibility of recruitment, and the basis for justifying the NI margin, alternative approaches using different NI margins and levels of alpha (such as 90% CIs) could be considered to make trials more feasible. We acknowledge that the use of 80% or 90% CIs increases the risk of falsely concluding noninferiority when compared with the traditional 95% CI, and this is the reason this approach should be reserved for situations in which new treatments are urgently needed for a small patient population where a trial using a 95% CI is not possible. In these situations, additional flexibility may be warranted and would maintain scientific credibility by explicitly stating the error risk levels and would be particularly justified in the situation we envisage where the trial results are also augmented with robust information from in vivo and in vitro models of infection. We feel that this is also aligned with the intention stated in FDA guidance, where FDA can approve if there is “a positive benefit-risk balance in the limited population…even though insufficient data exist to conclude that there is a favorable benefit-risk profile in a broader population.” Our calculation of patients benefitting from the chosen therapy assumed that all patients are either included in the trial or benefit from its results by receiving the preferred treatment. In reality, some patients will receive standard treatment outside the trial while it is ongoing. Including these patients in the calculations of the expected proportion of patients responding would lead to even smaller trials being optimal when the new agent was effective (as even more patients are receiving a suboptimal therapy) and would make little difference when it was not effective. Therefore, as the trial sample size becomes large, the expected number of patients benefitting from the chosen therapy would decrease even more rapidly than in the calculations presented in this paper. We acknowledge that this is a different level of evidence of efficacy than is usually used for regulatory approval, and this is why it is critical to gain the agreement of regulatory agencies and the infectious disease community when designing such studies that there is sufficient unmet and a small population which justifies this alternate approach to provide interpretable data for this patient population that would not otherwise be available. Furthermore, and as with any NI trial, it is important to be sure of the effectiveness of the control agent, and that the primary endpoint is objective and reflects a true clinical benefit to patients. Beyond a change in statistical criteria for the primary endpoint, agreement to such a design and any final assessment would always consider the benefit-risk profile of a new agent. This would include detailed review of secondary efficacy endpoints. Knowledge of safety risks of an agent before the trial, or with a specific drug class, may impact the acceptability of a smaller safety dataset. These considerations are specific to a given agent but would always need to be considered when designing a trial as we have proposed. This approach is relevant for drugs with activity only against a specific pathogen(s), and also for drugs with broad-spectrum activity which includes an uncommon pathogen. In both cases, this would provide important, quantitative data on pathogens of interest, with the key difference being the amount of additional information it is possible to generate.

CONCLUSIONS

In conclusion, registrational trials for rare pathogens are difficult, and sometimes impossible, to conduct using traditional statistical criteria. Explicit consideration by the medical, regulatory, and patient community of the ability to correctly approve an effective therapy, the risk of incorrectly approving an ineffective therapy, and the overall number of patients expected to benefit should be used to help define clear success criteria to ensure such studies can be conducted, so that data are available in these important patient groups.
  9 in total

1.  Clinical trials and rare diseases.

Authors:  Stephen W Lagakos
Journal:  N Engl J Med       Date:  2003-06-12       Impact factor: 91.245

2.  The value of information and optimal clinical trial design.

Authors:  Andrew R Willan; Eleanor M Pinto
Journal:  Stat Med       Date:  2005-06-30       Impact factor: 2.373

3.  Using value-of-information methods when the disease is rare and the treatment is expensive--the example of hemophilia A.

Authors:  Lusine Abrahamyan; Andrew R Willan; Joseph Beyene; Marjorie Mclimont; Victor Blanchette; Brian M Feldman
Journal:  J Gen Intern Med       Date:  2014-08       Impact factor: 5.128

4.  Plazomicin for Infections Caused by Carbapenem-Resistant Enterobacteriaceae.

Authors:  James A McKinnell; Jamie P Dwyer; George H Talbot; Lynn E Connolly; Ian Friedland; Alex Smith; Adrian M Jubb; Alisa W Serio; Kevin M Krause; George L Daikos
Journal:  N Engl J Med       Date:  2019-02-21       Impact factor: 91.245

5.  A comprehensive regulatory framework to address the unmet need for new antibacterial treatments.

Authors:  John H Rex; Barry I Eisenstein; Jeff Alder; Mark Goldberger; Robert Meyer; Aaron Dane; Ian Friedland; Charles Knirsch; Wendy R Sanhai; John Tomayko; Cindy Lancaster; Jennifer Jackson
Journal:  Lancet Infect Dis       Date:  2013-01-15       Impact factor: 25.071

6.  Approaches to sample size calculation for clinical trials in rare diseases.

Authors:  Frank Miller; Sarah Zohar; Nigel Stallard; Jason Madan; Martin Posch; Siew Wan Hee; Michael Pearce; Mårten Vågerö; Simon Day
Journal:  Pharm Stat       Date:  2018-01-10       Impact factor: 1.894

7.  Itraconazole to prevent fungal infections in chronic granulomatous disease.

Authors:  John I Gallin; David W Alling; Harry L Malech; Robert Wesley; Deloris Koziol; Beatriz Marciano; Eli M Eisenstein; Maria L Turner; Ellen S DeCarlo; Judith M Starling; Steven M Holland
Journal:  N Engl J Med       Date:  2003-06-12       Impact factor: 91.245

8.  Value of information methods to design a clinical trial in a small population to optimise a health economic utility function.

Authors:  Michael Pearce; Siew Wan Hee; Jason Madan; Martin Posch; Simon Day; Frank Miller; Sarah Zohar; Nigel Stallard
Journal:  BMC Med Res Methodol       Date:  2018-02-08       Impact factor: 4.615

Review 9.  Progress in the Fight Against Multidrug-Resistant Bacteria 2005-2016: Modern Noninferiority Trial Designs Enable Antibiotic Development in Advance of Epidemic Bacterial Resistance.

Authors:  John H Rex; George H Talbot; Mark J Goldberger; Barry I Eisenstein; Roger M Echols; John F Tomayko; Michael N Dudley; Aaron Dane
Journal:  Clin Infect Dis       Date:  2017-07-01       Impact factor: 9.079

  9 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.