Literature DB >> 35595961

Improvements to Survey Design from Pilot Testing a Discrete-Choice Experiment of the Preferences of Persons Living with HIV for Long-Acting Antiretroviral Therapies.

Douglas Barthold1, Aaron T Brah2, Susan M Graham2,3,4, Jane M Simoni3,5, Brett Hauber6,7.   

Abstract

BACKGROUND AND
OBJECTIVE: Development of clear and effective discrete-choice experiment surveys is an important step toward ensuring accurate and usable preference results. Pretest interviews and pilot testing are common in the development of discrete-choice experiments, and it is important for researchers to report details of survey changes resulting from patient feedback elicited in pilot work. This paper details pilot testing of an online discrete-choice experiment to elicit preferences for long-acting antiretroviral therapies among patients with HIV.
METHODS: The survey included an introduction to hypothetical treatment options, descriptions of attributes, comprehension questions, instructions for completing a discrete-choice experiment, a discrete-choice experiment with 17 choice tasks, and questions about personal characteristics. We piloted the survey with 50 respondents over ten waves. Each wave incorporated design improvements based on observations made during the previous wave. Respondents completed the online survey while screen sharing with a researcher, allowing interactive discussion. We developed a scheme for assessing and categorizing the survey changes.
RESULTS: Changes to the pilot were categorized by ways they impacted aspects of the discrete-choice experiment or the likely quality of resulting data. The four categories of impact are: understanding of attributes, underlying discrete-choice experiment and understanding of the choice question, collection of individual characteristics hypothesized to affect preference, and changes that improved clarity and usability of the survey without directly affecting the other categories (e.g., survey navigation and instructional clarity, formatting changes).
CONCLUSIONS: Detailed attention to the respondent experience in this large pilot allowed survey improvements that will likely reduce ambiguity, ensure more accurate capture of patient preferences and, ultimately, improve product development for long-acting antiretroviral therapies.
© 2022. The Author(s), under exclusive licence to Springer Nature Switzerland AG.

Entities:  

Mesh:

Year:  2022        PMID: 35595961      PMCID: PMC9123156          DOI: 10.1007/s40271-022-00581-z

Source DB:  PubMed          Journal:  Patient        ISSN: 1178-1653            Impact factor:   3.481


Key Points for Decision Makers

Introduction

Discrete-choice experiments (DCEs) are a valuable method for quantifying patient preferences for emerging health technologies [1]. Results of these studies allow product developers to align new product attributes with patient preferences, which improves acceptability of these new products. Most published reports of DCEs focus on study results, but it is also important for researchers to publish descriptions of their methodology and processes for building and executing DCEs. Such descriptions of formative qualitative research, pretest interviews, and pilot testing are important, and should be included in reports of DCEs in order to improve evaluation of survey performance in eliciting preferences and the interpretability of results. In this paper, we address this gap by describing the pilot testing of a DCE examining patient preferences for long-acting antiretroviral therapy (LA-ART) for HIV treatment, and documenting the changes that were made to the DCE survey as a result of pilot testing. Currently, most people living with HIV in the USA take one or more daily oral tablets combining two or more antiretroviral medications for treatment, the success of which is evaluated by viral suppression (i.e., an undetectable HIV viral load). Despite advances including the availability of single-tablet daily regimens for HIV, many patients still face challenges with antiretroviral therapy (ART) initiation and adherence [2]. Of the 1.1 million people in the USA living with HIV, approximately 86% have been diagnosed, 65% are receiving ART, and 56% are virally suppressed [3]. Long-acting formulations of antiretroviral medications, taken in combination, are an emerging HIV treatment modality with the potential to increase ART uptake and adherence [4]. New LA-ART modalities will provide increased options for patients who face barriers to taking daily oral pills, but the acceptability of these treatment regimens remains unclear. For this reason, our research team is conducting a DCE in the USA, which will establish the LA-ART product attributes and individual patient characteristics that will drive end user acceptability. Attributes and levels for a DCE designed to elicit preferences for potential LA-ART options were identified using a systematic process [5]. This process started with eight attributes defining potential LA-ART products, based on existing literature and knowledge of products in development. Then, we conducted 12 key informant interviews with HIV treatment experts. These interviews allowed iterative updating of the list of attributes, the set of plausible levels for each attribute, and restrictions on the combinations of attribute levels. Key informants converged on four delivery modes (long-acting oral tablets, subcutaneous injections, intramuscular injections, implants), and six other attributes of LA-ART: frequency of dosing, location of treatment, pain, pre-treatment time undetectable, pre-treatment negative reaction testing, and late-dose leeway [5]. The ISPOR checklist for conjoint analysis applications in health advises researchers to be fully transparent about their methodology and processes for study design and execution [6]. This paper documents the processes of pilot testing and refinement of our DCE of LA-ART acceptability, and fills the gap between attribute development and survey fielding.

Methods

Study Setting and Population

The pilot testing of this survey included ten waves of survey administration to 50 respondents and was conducted from September 2020 through February 2021. Respondents were recruited from the University of Washington (UW) HIV registry and directly from the UW HIV clinics in western Washington. The UW HIV Information System contains clinical information on patients with HIV seen since 1 January, 1995 at a network of HIV clinics run by the UW in Seattle and the surrounding region who have consented to the use of their de-identified data for research. Respondents were considered eligible for this study if they met the following inclusion criteria: (1) HIV-positive status, (2) enrollment in care at the UW Madison Clinic or one of its satellite clinics in western Washington, (3) age 18 years or older, (4) fluent in English, and (5) capable of providing informed consent. Additionally, respondents were excluded if they met the following exclusion criteria: (1) persons who are cognitively impaired or were under the influence of drugs or alcohol at the time of the pilot session; (2) persons who are “long-term non-progressors,” meaning they have maintained an undetectable viral load despite not taking ART; and (3) persons who are taking long-acting injectable ART already as part of a clinical trial. Because in-person participation was very limited owing to COVID-19, respondents also had to be able to use Zoom video teleconferencing and have an active e-mail address.

Pilot Process

The pilot testing occurred using Health Information Portability and Accountability Act-compliant Zoom (Zoom Video Communications, San Jose, CA, USA). Using this platform, the respondent completed the online survey using their own device (a computer, tablet, or smartphone), while screen sharing and discussing the survey with a member of the research team (ATB). As they completed the survey, respondents were encouraged to “think aloud,” and asked to describe their understanding of the product attributes and levels, as well as their understanding of the survey and DCE instructions. All pilot interviews were recorded on Zoom, with both audio and video of the screen sharing. After each respondent completed the pilot test, the researcher who conducted the pilot test (ATB) completed a qualitative assessment of the performance and comprehension of the respondent. This included a rating of whether and the extent to which the respondent required prompting to describe the choices in the DCE. Full details of this evaluation are included in the Appendix in the Electronic Supplementary Material (ESM).

Survey Instrument

The survey featured 88 questions, and was administered using SurveyEngine (SurveyEngine GmbH [7], Berlin, Germany), a web-based platform for administering DCE surveys. The first portion of the survey presented background information about the purpose of the study. We explained the details of the hypothetical modalities of LA-ART products, and the relevant attributes of these treatments. The presentation of this information featured comprehension questions to ensure respondent understanding of the content (see the ESM for the full text of these questions), and two “practice DCE” scenarios, where respondents viewed the information in the same format as the DCE. The survey also included a series of questions about the respondent’s history with HIV and HIV treatment; these characteristics are hypothesized to affect patient preferences for LA-ART. The next section was the DCE, which initially presented 16 choice scenarios to each respondent (a 17th scenario was later added). Every choice featured three alternatives, which varied across the seven product attributes. Two of the alternatives were hypothetical LA-ART regimens, and one alternative was the respondent’s current treatment (daily oral tablets). Respondents were randomly assigned to one of four blocks. Each block was associated with 16 out of 64 possible choice scenarios. The 16 scenarios for the respondent’s block were shown to the respondent in a random order. The DCE used an unlabeled experimental design developed with Ngene software (ChoiceMetrics, Sydney, NSW, Australia), with a modified Federov algorithm and D-error minimization for at least 3 minutes. The final choice sets had a D-error of 0.053978. Following the DCE, respondents were asked a series of questions about other personal characteristics that were hypothesized to affect LA-ART preferences. This includes access to healthcare, internalized HIV stigma, sociodemographic characteristics, sexuality, employment, and the experience of making choices in the DCE. Institutional review board approval for the entire study (including key informant interviews, pilot testing, and the full survey) was granted by the University of Washington Human Subjects Division (STUDY00007390). Respondents consented to participate both verbally and electronically at the time of pilot testing.

Pilot Feedback and Improvements

Recordings of the pilot tests were shared with members of the research team in a Health Information Portability and Accountability Act-compliant cloud storage drive. The study team evaluated these recordings for points of confusion or any other potential area of improvement for the survey. Each week, the team discussed these potential improvements during a regular meeting, and iteratively incorporated changes into subsequent waves of the pilot. All changes to the survey were systematically logged. Successive waves had the following number of unique respondents: wave 1 (8 respondents), 2 (12), 3 (6), 4 (1), 5 (4), 6 (3), 7 (4), 8 (1), 9 (3), 10 (8). The number of respondents varied across wave owing to the iterative nature of the survey improvements and to differences in the time needed for survey edits and recruitment across waves. We ended with ten waves, as this was the number of iterations that were undertaken before completing the interview with the 50th respondent. Upon completion of the pilot testing, every change to the survey from the pilot log was categorized according to its impact on the DCE and the quality of the resulting data. Substantive changes were those that we deemed to have probable effects on the DCE and survey data collection, and these changes converged into three major categories that were identified post-hoc: (1) changes that likely impacted the understanding of the attributes, (2) changes that likely impacted the underlying DCE design and understanding of the choice question, and (3) changes that likely impacted the collection of individual characteristics that may affect choice. Finally, there was a fourth category that included the changes unlikely to be substantive; these were changes that improved survey navigation and instructional clarity, clarified the meaning of questions and their options, or adjusted formatting. After all changes were made, the full survey (ESM) included 88 questions (five comprehension questions, two practice DCE scenarios, 64 covariates, and 17 DCE scenarios).

Data Analysis

Summary statistics were used to present characteristics of the pilot test respondent sample. We also used summary statistics to describe the respondents’ performance on the comprehension questions across waves (proportion correct on first try, and proportion ever correct), and the proportion of respondents who were able to describe the treatment choices in the practice DCEs without prompting.

Results

Characteristics of the pilot respondents in all waves are displayed in Table 1. From wave 1 to wave 10, the proportion of respondents who required prompting to describe the choices in the first practice DCE decreased from 63 to 0%, and the proportion who required prompting to describe the choices in the second practice DCE decreased from 38 to 0% (Table 2). However, the comprehension questions completed by respondents, which were true/false questions about each attribute, showed no meaningful change in performance over the course of the pilot waves (Table 2). Table 1 also shows that among scenarios with long-acting oral included as an option, 61% of choices were for long-acting oral (the same figure was 54% for injections, and 47% for implants). Current therapy was an option in every scenario, and 17% of choices were for this option.
Table 1

Description of pilot test respondents

N (%) or median (min, IQR, max)
Gender identity
 Female gender6 (12%)
 Male gender43 (86%)
 Trans man0 (0%)
 Trans woman1 (2%)
Sexual orientation
 Gay or lesbian35 (71%)
 Bisexual5 (10%)
 Straight9 (18%)
Race/ethnicity
 Black9 (18%)
 Hispanic8 (16%)
 White33 (66%)
 Other7 (14%)
Age52 (23, 42–58, 68)
Current HIV treatment regimen
 1 pill per day25 (50%)
 2 pills per day19 (38%)
 3+ pills per day5 (10%)
Years since ART start15 (1, 11–25, 38)
Years since HIV diagnosis20 (1, 12–30, 41)
Education: high school or less18 (36%)
Education: some college17 (34%)
Education: college graduate or graduate degree15 (30%)
Survey choices
 Always chose current therapy0 (0%)
 Ever chose long-acting oral45 (90%)
 Ever chose an injection46 (92%)
 Ever chose an implant29 (58%)
 Percent of scenarios chose current therapy17%
 Percent of scenarios chose LAO61%
 Percent of scenarios chose injection54%
 Percent of scenarios chose implant47%
Survey duration in minutes47 (25, 37–62, 132)

Sample is 50 respondents in a pilot test of a survey and a discrete-choice experiment of patient preferences for long-acting antiretrovirals. Sexual orientation does not sum to 50 because one respondent selected “other”; race/ethnicity does not sum to 50 because some people selected multiple options; current HIV treatment regimen does not sum to 50 because one respondent indicated they had never started treatment. Some college refers to “Some college/AA degree/Technical school training”. The percentages of scenarios are among scenarios where that mode was an option

ART antiretroviral therapy, IQR interquartile range, LAO long-acting oral, max maximum, min minimum

Table 2

Pilot test respondent performance across waves of the pilot

All wavesW1W2W3W3W5W6W7W8W9W10
(n = 50)(n = 8)(n = 12)(n = 6)(n = 1)(n = 4)(n = 3)(n = 4)(n = 1)(n = 3)(n = 8)
Comprehension question performance
 Q1 correct on first try76%88%83%67%100%50%33%100%100%67%75%
 Q1 ever correct98%100%100%100%100%75%100%100%100%100%100%
 Q2 correct on first try86%88%100%50%100%75%67%75%100%100%100%
 Q2 ever correct100%100%100%100%100%100%100%100%100%100%100%
 Q3 correct on first try98%100%100%100%100%100%100%75%100%100%100%
 Q3 ever correct100%100%100%100%100%100%100%100%100%100%100%
 Q4 correct on first try86%63%83%83%100%100%100%100%100%100%88%
 Q4 ever correct100%100%100%100%100%100%100%100%100%100%100%
 Q5 correct on first try92%88%92%100%100%75%100%100%0%100%100%
 Q5 ever correct100%100%100%100%100%100%100%100%100%100%100%
Qualitative assessments of respondents’ understanding of the practice DCEs
 Struggled to describe options in first practice DCE28%63%50%17%0%25%0%0%0%33%0%
 Struggled to describe options in second practice DCE18%38%25%0%0%25%0%0%0%67%0%

Comprehension question performance depicts percent of respondents who correctly answered each true/false comprehension question. Respondents had two opportunities to answer each question. Qualitative assessments depict percent of respondents who required prompting in order to describe the three-choice options in the two practice DCEs (see Appendix in the ESM for details on this measure). Sample is 50 respondents in a pilot test of a survey and a discrete-choice experiment of patient preferences for long-acting antiretrovirals

DCE discrete-choice experiment, Q1 comprehension question 1, W1 pilot wave 1

Description of pilot test respondents Sample is 50 respondents in a pilot test of a survey and a discrete-choice experiment of patient preferences for long-acting antiretrovirals. Sexual orientation does not sum to 50 because one respondent selected “other”; race/ethnicity does not sum to 50 because some people selected multiple options; current HIV treatment regimen does not sum to 50 because one respondent indicated they had never started treatment. Some college refers to “Some college/AA degree/Technical school training”. The percentages of scenarios are among scenarios where that mode was an option ART antiretroviral therapy, IQR interquartile range, LAO long-acting oral, max maximum, min minimum Pilot test respondent performance across waves of the pilot Comprehension question performance depicts percent of respondents who correctly answered each true/false comprehension question. Respondents had two opportunities to answer each question. Qualitative assessments depict percent of respondents who required prompting in order to describe the three-choice options in the two practice DCEs (see Appendix in the ESM for details on this measure). Sample is 50 respondents in a pilot test of a survey and a discrete-choice experiment of patient preferences for long-acting antiretrovirals DCE discrete-choice experiment, Q1 comprehension question 1, W1 pilot wave 1 The full list of changes that occurred during the pilot testing, within their respective categories, are displayed in Fig. 1, with annotations to indicate in which wave the change was made. Figure 1a shows the changes that were made to improve the respondents’ understanding of the attributes of the hypothetical LA-ART products. All attributes received specific changes, except frequency of dosing and pain, which were well understood by all respondents from the beginning. We added clickable icons within the DCE, which show the descriptions of each attribute, using the exact same wording that was used during the section of the survey that first presented the attribute description. Another important change was the wording of the concept of “time undetectable,” which refers to the amount of time a respondent would need an undetectable viral load prior to initiating a hypothetical LA-ART product. Initially, this undetectability was referred to as having achieved “viral suppression,” but we found that not all respondents were familiar with this terminology. Being “undetectable” is more commonly discussed by respondents with their healthcare providers, and we therefore changed this wording to “pre-treatment time undetectable.” We also expanded the content that describes the “late-dose leeway” attribute, to add examples of short and long leeway time periods, and used precise wording to explain that if doses are received outside the leeway period, then the HIV viral load can become detectable.
Fig. 1

a Changes that likely impact the understanding of the attributes, across waves. b Changes that likely impact the underlying discrete-choice experiment design, and the understanding of the choice question, across waves. c Changes that likely impact the collection of individual characteristics that may affect choice, across waves. W wave

a Changes that likely impact the understanding of the attributes, across waves. b Changes that likely impact the underlying discrete-choice experiment design, and the understanding of the choice question, across waves. c Changes that likely impact the collection of individual characteristics that may affect choice, across waves. W wave Figure 1b lists the changes that impacted the underlying DCE design and the understanding of the choice question. This included adding emphasis that each choice was between two hypothetical long-acting treatment options and current therapy (daily oral tablets). In the early waves of the pilot, some respondents struggled to understand how to interpret and complete the practice DCE scenarios that are shown early in the survey. To alleviate this problem, we added two videos that explained the elements of the DCE choice scenario (see Fig. 2 for a screenshot of the second video, and the ESM for the full videos). In the videos, a voiceover from a research team member described the details of all three alternatives (displayed in vertical columns), including the level of each attribute for that alternative, with red arrows to guide the respondent’s attention as each row within each column was described. The researcher who conducted the interviews (ATB) noted that the sequence of two separate videos was helpful: the first video featured a practice DCE with only four relatively simple attributes (mode, location, frequency, and pain), and allowed respondents to become oriented with the structure of the choice sets. Then, the second video added the more complex attributes. This sequence helped certain respondents to more fully appreciate the variations between choice sets. Another important change was the addition of a 17th scenario to the DCE, in which the respondent chose between long-acting oral options and their current treatment. This scenario was added in order to observe the trade-offs that people make among the other attributes (aside from treatment type), within the long-acting oral modality, which we thought might be more popular than other treatment modalities.
Fig. 2

Screenshot of video that explains the content on the discrete-choice experiment choice screen

Screenshot of video that explains the content on the discrete-choice experiment choice screen Figure 1c shows the changes that were related to the collection of personal characteristics that we hypothesize could be related to treatment choice preferences. In the first nine waves of the survey, we used the 12-item version of the Berger Stigma Scale, and this was emotionally burdensome on many respondents [8]. For the tenth wave, we switched to a shorter questionnaire, which reduced this burden [9]. We also added two questions to the end of the survey, which inquire about the respondent’s decision-making process in the DCE. This was intended to identify respondents who did not consider all attributes of the alternatives when making their choice. In the figure in the ESM, we show the full list of all other changes that were made to the survey during the pilot testing. These changes were important for improving the clarity and usability of the survey, but did not directly affect the DCE attributes, choice question, or collection of individual characteristics. These changes were made in order to improve survey navigation and instructional clarity, adjust formatting, and clarify the meaning of questions and their options.

Discussion

This paper describes the pilot testing of a DCE that will examine patient preferences for LA-ART, which are important emerging therapies for the treatment of HIV. We used a systematic and standardized process for conducting and recording the details of each respondent’s experience, therefore allowing our research team to identify the parts of the survey that were not clear to respondents or could be improved. Our process included both qualitative and quantitative assessments of the respondents’ experience. In a relatively large sample of respondents (n = 50), we made iterative improvements to the survey over ten waves. These changes clarified points of confusion regarding the hypothetical LA-ART products, as well as confusion in the process of completing a DCE. During our pilot testing, we found that respondents understood our instructions for the survey quite well, especially after we added videos that showed examples of how to review, consider, and complete the choice tasks. The respondents showed good understanding of most of the attributes. Pain, frequency, and product type were well understood by respondents throughout. Even pre-treatment negative reaction testing, which we assumed was a more complicated concept, was well understood from the start. The more challenging concepts were pre-treatment time undetectable and late-dose leeway, which are important attributes of the first LA-ART regimen on the market, Cabenuva (cabotegravir-rilpivirine). The two key clinical trials demonstrating efficacy of this regimen required participants to have an undetectable viral load (<50 copies/mL) on their current HIV regimen for at least 6 months [10], or after 16 weeks on an oral regimen of dolutegravir-abacavir-lamivudine [11]. Moreover, because of concerns about the emergence of antiretroviral resistance if doses are delayed, detailed guidelines have been developed for close patient management by providers and clinics [12]. The resulting alterations in survey content and clarity were aimed to enhance accurate capture of patient preferences and characteristics, which can reduce ambiguity and potentially increase precision [13]. These changes increased our confidence that the survey is measuring what we intend to measure, and that it will ultimately establish how certain patient characteristics are associated with the acceptability of the tested attributes. This process will therefore facilitate maximum influence of the data that results from our DCE, and ultimately improve understanding of product development needs with respect to LA-ART, with long-term implications for patient acceptability, adherence, and health. Strengths of our study include the large sample of pilot respondents, the recordings and specific attention to each respondent’s experience with the survey, and systematic logging of all changes made to the survey. One limitation of the study is that despite our efforts, it is possible that some feedback from respondents was overlooked or missed because we did not conduct in-depth exit interviews. Additionally, there is not a direct link between every piece of information gained and specific changes to the survey. Another limitation is that because of a requirement for online pilot testing owing to COVID-19, all of our pilot respondents had sufficient computer literacy to complete the survey on Zoom while screen sharing with a member of our research team; this may have inadvertently led to the selection of respondents with higher education and better computer literacy, who may not be representative of the general population of people living with HIV. Similarly, our recruitment, which occurred strictly within western Washington, reduces the representativeness of the sample. Finally, while we believe that these changes improved the survey and will lead to more precise data collection, we do not have proof of the connection between the survey changes and the quality of the resulting data.

Conclusions

Reporting on the pilot testing of this DCE survey provides transparency in our methodology and processes for study design and execution, which we hope will improve the interpretability of our findings. While details of formative qualitative research to inform DCE design are sometimes published [14], the processes of pilot testing are not. We recommend that for all DCEs, the process of testing the survey instrument, and the details of changes that may directly impact the survey’s data collection, be reported somewhere easily accessible to readers of the results, either as supplementary material or in a stand-alone article. This is the first paper, to our knowledge, to fully report on these practices. This reporting demonstrates that the survey worked as intended, and measured what it intended to measure, which increases the credibility of the results. Below is the link to the electronic supplementary material. Supplementary file1 (PDF 209 KB) Supplementary file2 (PDF 900 KB) Supplementary file3 (MOV 10197 KB) Supplementary file4 (MOV 16733 KB)
This paper details the pilot testing process for a discrete-choice experiment that examines patient preferences for long-acting antiretroviral therapies, and summarizes revisions to the survey that improved respondent understanding of the attributes, clarified the discrete-choice experiment choice questions and design, and optimized collection of individual characteristics that may affect choice.
Pretest interviews and pilot testing are common in the development of discrete-choice experiments, and it is important for researchers to report details of survey changes resulting from patient feedback elicited in pilot work.
Detailed attention to the respondent experience in this large pilot allowed survey improvements, such as the addition of visual aids, that will likely reduce ambiguity, ensure more accurate capture of patient preferences, and provide an example for other researchers to follow in conducting and describing qualitative work that supports the development of discrete-choice experiments.
  12 in total

1.  Conjoint analysis applications in health--a checklist: a report of the ISPOR Good Research Practices for Conjoint Analysis Task Force.

Authors:  John F P Bridges; A Brett Hauber; Deborah Marshall; Andrew Lloyd; Lisa A Prosser; Dean A Regier; F Reed Johnson; Josephine Mauskopf
Journal:  Value Health       Date:  2011-04-22       Impact factor: 5.725

2.  Understanding the HIV care continuum.

Authors:  Robert S Hogg
Journal:  Lancet HIV       Date:  2018-06       Impact factor: 12.767

3.  Long-Acting Cabotegravir and Rilpivirine for Maintenance of HIV-1 Suppression.

Authors:  Susan Swindells; Jaime-Federico Andrade-Villanueva; Gary J Richmond; Giuliano Rizzardini; Axel Baumgarten; Mar Masiá; Gulam Latiff; Vadim Pokrovsky; Fritz Bredeek; Graham Smith; Pedro Cahn; Yeon-Sook Kim; Susan L Ford; Christine L Talarico; Parul Patel; Vasiliki Chounta; Herta Crauwels; Wim Parys; Simon Vanveggel; Joseph Mrus; Jenny Huang; Conn M Harrington; Krischan J Hudson; David A Margolis; Kimberly Y Smith; Peter E Williams; William R Spreen
Journal:  N Engl J Med       Date:  2020-03-04       Impact factor: 91.245

4.  Long-Acting Cabotegravir and Rilpivirine after Oral Induction for HIV-1 Infection.

Authors:  Chloe Orkin; Keikawus Arasteh; M Górgolas Hernández-Mora; Vadim Pokrovsky; Edgar T Overton; Pierre-Marie Girard; Shinichi Oka; Sharon Walmsley; Chris Bettacchi; Cynthia Brinson; Patrick Philibert; Johan Lombaard; Marty St Clair; Herta Crauwels; Susan L Ford; Parul Patel; Vasiliki Chounta; Ronald D'Amico; Simon Vanveggel; David Dorey; Amy Cutrell; Sandy Griffith; David A Margolis; Peter E Williams; Wim Parys; Kimberly Y Smith; William R Spreen
Journal:  N Engl J Med       Date:  2020-03-04       Impact factor: 91.245

5.  Reporting Formative Qualitative Research to Support the Development of Quantitative Preference Study Protocols and Corresponding Survey Instruments: Guidelines for Authors and Reviewers.

Authors:  Ilene L Hollin; Benjamin M Craig; Joanna Coast; Kathleen Beusterien; Caroline Vass; Rachael DiSantostefano; Holly Peay
Journal:  Patient       Date:  2020-02       Impact factor: 3.883

6.  Constructing experimental designs for discrete-choice experiments: report of the ISPOR Conjoint Analysis Experimental Design Good Research Practices Task Force.

Authors:  F Reed Johnson; Emily Lancsar; Deborah Marshall; Vikram Kilambi; Axel Mühlbacher; Dean A Regier; Brian W Bresnahan; Barbara Kanninen; John F P Bridges
Journal:  Value Health       Date:  2013 Jan-Feb       Impact factor: 5.725

7.  Development of a 12-item short version of the HIV stigma scale.

Authors:  Maria Reinius; Lena Wettergren; Maria Wiklander; Veronica Svedhem; Anna Mia Ekström; Lars E Eriksson
Journal:  Health Qual Life Outcomes       Date:  2017-05-30       Impact factor: 3.186

8.  HIV virologic response better with single-tablet once daily regimens compared to multiple-tablet daily regimens.

Authors:  Shashi N Kapadia; Robert R Grant; Susan B German; Baljinder Singh; Amy L Davidow; Shobha Swaminathan; Sally Hodder
Journal:  SAGE Open Med       Date:  2018-12-04

9.  Discrete Choice Experiments in Health Economics: Past, Present and Future.

Authors:  Vikas Soekhai; Esther W de Bekker-Grob; Alan R Ellis; Caroline M Vass
Journal:  Pharmacoeconomics       Date:  2019-02       Impact factor: 4.981

10.  Real-world adherence and persistence for newly-prescribed HIV treatment: single versus multiple tablet regimen comparison among US medicaid beneficiaries.

Authors:  Joshua Cohen; Anne Beaubrun; Richa Bashyal; Ahong Huang; Jieni Li; Onur Baser
Journal:  AIDS Res Ther       Date:  2020-04-01       Impact factor: 2.250

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.