Literature DB >> 29581091

How Online Quality Ratings Influence Patients' Choice of Medical Providers: Controlled Experimental Survey Study.

Niam Yaraghi1,2, Weiguang Wang3, Guodong Gordon Gao3, Ritu Agarwal3.   

Abstract

BACKGROUND: In recent years, the information environment for patients to learn about physician quality is being rapidly changed by Web-based ratings from both commercial and government efforts. However, little is known about how various types of Web-based ratings affect individuals' choice of physicians.
OBJECTIVE: The objective of this research was to measure the relative importance of Web-based quality ratings from governmental and commercial agencies on individuals' choice of primary care physicians.
METHODS: In a choice-based conjoint experiment conducted on a sample of 1000 Amazon Mechanical Turk users in October 2016, individuals were asked to choose their preferred primary care physician from pairs of physicians with different ratings in clinical and nonclinical aspects of care provided by governmental and commercial agencies.
RESULTS: The relative log odds of choosing a physician increases by 1.31 (95% CI 1.26-1.37; P<.001) and 1.32 (95% CI 1.27-1.39; P<.001) units when the government clinical ratings and commercial nonclinical ratings move from 2 to 4 stars, respectively. The relative log odds of choosing a physician increases by 1.12 (95% CI 1.07-1.18; P<.001) units when the commercial clinical ratings move from 2 to 4 stars. The relative log odds of selecting a physician with 4 stars in nonclinical ratings provided by the government is 1.03 (95% CI 0.98-1.09; P<.001) units higher than a physician with 2 stars in this rating. The log odds of selecting a physician with 4 stars in nonclinical government ratings relative to a physician with 2 stars is 0.23 (95% CI 0.13-0.33; P<.001) units higher for females compared with males. Similar star increase in nonclinical commercial ratings increases the relative log odds of selecting the physician by female respondents by 0.15 (95% CI 0.04-0.26; P=.006) units.
CONCLUSIONS: Individuals perceive nonclinical ratings provided by commercial websites as important as clinical ratings provided by government websites when choosing a primary care physician. There are significant gender differences in how the ratings are used. More research is needed on whether patients are making the best use of different types of ratings, as well as the optimal allocation of resources in improving physician ratings from the government's perspective. ©Niam Yaraghi, Weiguang Wang, Guodong (Gordon) Gao, Ritu Agarwal. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 26.03.2018.

Entities:  

Keywords:  health care evaluation mechanisms; quality of health care

Mesh:

Year:  2018        PMID: 29581091      PMCID: PMC5891665          DOI: 10.2196/jmir.8986

Source DB:  PubMed          Journal:  J Med Internet Res        ISSN: 1438-8871            Impact factor:   5.428


Introduction

To improve quality, foster competition, promote transparency, and help patients make informed decisions, it is critical for patients to have access to reliable information and make cognizant choices about their medical providers [1,2]. In recent years, a concerted effort in the United States has been put in place to develop and publicly report quality measures of medical providers [3]. The Centers for Medicare and Medicaid Services (CMS) is the most prominent governmental agency in the United States that collects, aggregates, and reports quality measures of different aspects of medical care. Through initiatives such as Hospital Compare [4], CMS reports quality data on both clinical and nonclinical aspects of medical services offered by different providers. Surgical complications, infections, readmission, and death rates are examples of metrics that measure the clinical aspects of medical care. Surveys of patients’ experiences, such as the Hospital Consumer Assessment of Healthcare Providers and Systems, capture metrics that measure nonclinical aspects of care. In parallel with CMS, private and commercial agencies such as Vitals [5], RateMDs [6], and ProPublica [7] also collect and report quality metrics on both clinical and nonclinical aspects of care. Recent research shows that although the ratings provided by commercial agencies may be inconsistent with each other [8], they are more comprehensive and cover a broader range of domains than what is included in ratings reported by CMS [9,10]. Ratings of health care providers are growing in importance and popularity [11-18], affecting both the revenue and the reputation of medical providers [19-22]. For example, when CMS released its quality metrics of nursing homes to the public, the market share of 1-star facilities decreased by 8%, whereas the market share of 5-star facilities increased by more than 6% [23]. Similar effects have also been documented for hospitals [24]. Although nonclinical ratings provided by commercial agencies are correlated with the conventional measures of patient experience as reported by governmental agencies [25,26], the relationship between patient reviews and medical outcomes is not clear. Some studies find that patient satisfaction reported as nonclinical ratings is not associated with clinical outcomes [27-32], whereas others report a strong association between these two types of ratings [33,34]. For a review of literature on the association between the social media reviews and the clinical quality outcomes, see Verhoef et al [35]. Despite the significant differences between the types (clinical and nonclinical) and the sources (governmental and commercial agencies) of ratings, variations in their relative significance for patient choice of medical providers are not known. The purpose of this research was to fill this gap by uncovering the relative importance of these ratings in the decision-making processes of different groups of patients.

Methods

Data Source

We used a primary dataset consisting of responses of 1000 individuals who were each paid 50 cents to participate in an online experiment through Amazon Mechanical Turk (AMT) in October 2016. These individuals were all master users of AMT and live in the United States. According to AMT, a user achieves a master distinction by consistently completing requests with a high degree of accuracy. Masters must continue to pass AMT’s statistical monitoring to maintain their status [36]. Table 1 provides a comparison of demographics between the sample in this study and the US population. In contrast to the US population, our sample consisted of less affluent, but more educated, younger adults. Although, when compared with the US population, our sample of AMT users consisted of younger and more technologically savvy individuals, we relied on this sample to conduct our analysis for the following reasons. First, given the question posed in this research, the sample did not need to be representative of the US population and, instead, only had to represent individuals who used information resources available on the Internet. As this study compared the importance of two information resources that are exclusively Web-based, its sample also had to include the individuals who could use resources on the Web. Second, prior research shows that despite limitations, data that are gathered from “AMT samples are at least as reliable as those obtained via traditional methods. Overall, AMT can be used to obtain high-quality data inexpensively and rapidly” [37].
Table 1

Characteristics of 949 respondents and the US population.

Variable and classSample, n (%)Percentage of US populationa (%)H0: PUSAPSample=0b (z value)
Education
Advanced degree114 (12.0)10.38−1.65c
Bachelor’s degree381 (40.2)18.88−16.74d
Associate’s degree100 (10.5)5.28−7.25d
Some college, no degree231 (24.3)19.42−3.83d
Trade or technical school30 (3.2)4.081.43
Graduated high school90 (9.5)29.6313.59d
Less than high school3 (0.3)12.3311.25d
Income, US $
150,000 or more30 (3.2)13.579.36d
125,000-149,99925 (2.6)5.423.8d
100,000-124,99970 (7.4)8.711.45
75,000-99,999123 (13.0)12.26−0.66
50,000-74,999227 (23.9)16.96−5.71d
35,000-49,999170 (17.9)12.92−4.58d
25,000-34,999136 (14.3)9.39−5.22d
Less than 25,000168 (17.7)20.772.33e
Race
Asian56 (5.9)5.70−0.27
Black66 (7.0)13.305.76d
Hawaiian1 (0.1)0.200.62
Hispanic53 (6.0)17.89.84d
Indian16 (1.7)1.30−1.06
White757 (79.8)76.90−2.1e
Marital status
Divorced74 (7.8)9.802.07e
Married/Domestic partner471(49.6)51.871.38
Separated8 (0.8)2.092.69d
Single/Never married390 (41.1)32.25−5.83d
Widowed6 (0.6)5.726.75d
Gender
Female548 (57.7)50.80−4.28d
Male401 (42.3)49.204.28d
Age
Younger than 65 years924 (97.4)87.00−9.49d
65 years and older25 (2.6)13.009.49d

aAuthors’ analysis of characteristics of experiment participants. Demographics of US population are calculated based on the data provided by the US Census Bureau.

bThe null hypothesis that the percentage in sample is equal to that of the US population.

cP<.10.

dP<.01.

eP<.05.

Study Design

To determine how ratings on different attributes affect individuals’ evaluations of medical providers, we designed an experiment and conducted a choice-based conjoint analysis [38] as a rigorous method of eliciting preferences [39]. We describe the method below. The combination of 2 categories (clinical and nonclinical) and 2 sources (governmental and commercial agencies) resulted in 4 different types of ratings: clinical ratings provided by a governmental agency, nonclinical ratings provided by a governmental agency, clinical ratings provided by a commercial agency, and nonclinical ratings provided by a commercial agency. In this research, we use “governmental agency” and “public agency” interchangeably. We assigned a high or low value to each type of rating, and thereby created 16 profiles of hypothetical physicians. In a 1-to-4-star rating system, to induce appropriate variation, we used 2 stars to indicate low ratings and 4 stars to indicate high ratings. Each profile represented a physician with different ratings on the 4 categories. These profiles were balanced, which means that each of the 2 levels (2 and 4 stars) in each of the 4 types of ratings appeared the same number of times in physician profiles. Using these 16 profiles, we then created 8 pairs of physicians such that the 4 types of ratings in each pair were orthogonal [40]. This ensured that any pair of levels from different rating types appeared the same number of times in the design. We used % mktex [41] macro in SAS software (version 9.4) to create the balanced and orthogonal design. Table 2 shows the 16 profiles in 8 pairs.
Table 2

Physician profiles used in choice-based conjoint experiment. “gGvernment” indicates that a public agency provides the ratings, and “Commercial” indicates that a private organization provides the ratings. In the Web-based interface, the hypothetical physician profiles in each pair were shown side-by-side and respondents were asked to choose the physician they prefer. The sequence of the pairs and the attributes in each profile were generated randomly to ensure that the order of the presentation of rank of the attributes did not influence the respondent’s choice. The values of 2 or 4 in the table, respectively, indicate a “2” or “4” star rating in the physician profiles provided to respondents in the Web-based experiment.

Pair numberaGovernment ratingCommercial rating
ClinicalNonclinicalClinicalNonclinical
One2; 44; 22; 44; 2
Two2; 44; 24; 24; 2
Three2; 42; 22; 24; 2
Four4; 22; 42; 44; 2
Five4; 22; 44; 24; 2
Six4; 24; 22; 44; 2
Seven2; 42; 44; 24; 2
Eight2; 42; 42; 42; 4
In a Web-based interface, we first provided respondents with a brief tutorial on different sources and types of ratings. Specifically, we described the public agency as “the department of Health and Human Services, which is a branch of the federal government” and the commercial agency as “websites such as Yelp, RateMDs, Healthgrades, Vitals, Zocdoc, and DoctorScorecard.” Characteristics of 949 respondents and the US population. aAuthors’ analysis of characteristics of experiment participants. Demographics of US population are calculated based on the data provided by the US Census Bureau. bThe null hypothesis that the percentage in sample is equal to that of the US population. cP<.10. dP<.01. eP<.05. Physician profiles used in choice-based conjoint experiment. “gGvernment” indicates that a public agency provides the ratings, and “Commercial” indicates that a private organization provides the ratings. In the Web-based interface, the hypothetical physician profiles in each pair were shown side-by-side and respondents were asked to choose the physician they prefer. The sequence of the pairs and the attributes in each profile were generated randomly to ensure that the order of the presentation of rank of the attributes did not influence the respondent’s choice. The values of 2 or 4 in the table, respectively, indicate a “2” or “4” star rating in the physician profiles provided to respondents in the Web-based experiment. Screenshot of the choice-based conjoint experiment. We also distinguished clinical and nonclinical ratings and explained to the survey respondents that clinical ratings by the public agency were determined “based on official statistics on how often physicians provide care that research shows leads to the best results for patients” and nonclinical ratings by the public agency were determined based on “a national survey that asks patients about their experiences with staff, nurses, and doctors during a recent visit to the doctor.” Similarly, we explained that clinical ratings provided by the commercial agency were determined by “the patient online reviews about how patients evaluate the medical expertise of the doctor” and nonclinical ratings provided by the commercial agency were created based on “patient online reviews about their experiences with staff, nurses, and doctors during a recent visit to the doctor.” To assess if respondents correctly distinguished the differences between the types and the sources of ratings, at the end of the survey, we asked them to describe each type of the ratings in their own words. Our examination of their responses confirmed that all respondents had fully understood different ratings. We then presented the 8 pairs of hypothetical profiles of physicians in a random sequence and asked respondents to choose the physician they prefer in each pair. A screenshot of 1 of the 8 comparison pairs is presented in Figure 1, which corresponds to the choices in pair Seven as shown in Table 1. To simulate a realistic decision-making scenario, we asked the respondents to imagine that they have moved to a new town and have to choose a new primary care physician based solely on the 4 types of ratings provided to them. This approach ensured that the choice of the respondents in our experiment was only driven by the ratings and was not confounded by any other factor outside of our model, such as insurance coverage, location, or race of the physician [42,43].
Figure 1

Screenshot of the choice-based conjoint experiment.

Once respondents finished the evaluation of physicians in the 8 pairs, we asked them a series of questions designed to evaluate their health status, medical literacy, trust in Web-based reviews, and trust in government as 4 composite indexes. We conducted factor analysis to operationalize these 4 constructs using validated items that we derived from prior literature in information systems [44,45] and medicine [46,47]. Details on the items, composite indexes, and factor analysis are provided in Multimedia Appendix 1. One potential concern with the study design was that respondents may not complete the choice task thoughtfully. To detect and filter the responses that were provided hastily and without careful attention, we included 2 trap questions in the experiment. The first trap question was the choice of physicians in the eighth pair (shown in Table 2), one of which was superior on all of the 4 types of ratings and clearly dominated the pair. A respondent’s choice of an inferior physician indicated lack of attention to the experiment. The second trap question asked, “How happy will you be if you receive a letter from Internal Revenue Service that says you should pay a large amount of taxes to the government?” We assumed that a respondent did not pay attention to the question if she chose “extremely happy” or “happy” as a response to this question.

Statistical Analysis

Our research design fit the multinomial logit model with clustered error terms [48,49]. Following the suggestions of Kuhfeld [50], we used the PHREG [51] procedure in SAS software for the estimation. In this model, the dependent variable was binary and indicated the choice that a respondent made from a pair of hypothetical physician profiles. The 4 types of ratings in each profile constituted our main independent variables. In the multinomial logit model used in this study, the probability that a respondent chose a specific physician in a pair was a function of the attributes of that specific physicians as well as the attributes of the other physician in the pair. The PHREG [51] procedure in SAS not only allowed us to account for the conditional dependency of choices for the alternatives in a pair but also adjusted for the correlation between the 8 choices made by the same respondent. Using this model, we could examine the relative importance of the 4 types of ratings. We further explored whether patient attributes, such as age, gender, and income, moderated the impact of the ratings. To statistically compare the effects of different regression coefficients, we implemented the tests provided by Paternoster et al [52].

Results

On the basis of the answers to the 2 trap questions, we excluded 51 observations from our initial sample of 1000 responses. We retained the remaining 949 responses for further analysis (Table 1). We present the estimation results of our multinomial logit model in Table 3.
Table 3

The relative importance of different types and sources of ratings on patients’ choice. GC: clinical ratings provided by a public agency (government). GNC: nonclinical ratings provided by a public agency (government). YC: clinical ratings provided by a commercial agency (commercial). YNC: nonclinical ratings provided by a commercial agency (commercial).

ParameteraParameter estimate (95% CI)
Basic modelHealth statusMedical literacyTrust in online reviewsTrust in governmentFull model
GC1.29b (1.24 to 1.34)1.29b (1.24 to 1.35)1.29b (1.24 to 1.34)1.29b (1.24 to 1.34)1.30b (1.25 to 1.35)1.31b (1.26 to 1.36)
GNC1.00b (0.95 to 1.05)1.01b (0.96 to 1.06)1.00b (0.95 to 1.05)1.00b (0.95 to 1.05)1.01b (0.96 to 1.06)1.03b (0.98 to 1.08)
YC1.09b (1.04 to 1.14)1.11b (1.06 to 1.16)1.10b (1.04 to 1.14)1.10b (1.04 to 1.14)1.11b (1.07 to 1.16)1.12b (1.07 to 1.18)
YNC1.29b (1.24 to 1.34)1.31b (1.25 to 1.36)1.29b (1.24 to 1.34)1.29b (1.24 to 1.35)1.30b (1.25 to 1.35)1.32b (1.27 to 1.37)
Health status × GC−0.13c (−0.18 to −0.08)−0.13c (−0.19 to −0.08)
Health status × GNC0.09c (0.04 to 0.14)0.10c (0.05 to 0.15)
Health status × YC0.05 (0 to 0.10)0.05 (0 to 0.10)
Health status × YNC0.17b (0.12 to 0.22)0.18b (0.13 to 0.23)
Medical literacy × GC0 (−0.06 to 0.04)0 (−0.06 to 0.04)
Medical literacy × GNC0 (−0.05 to 0.04)−0.01 (−0.05 to 0.04)
Medical literacy × YC0.03 (−0.01 to 0.08)0.03 (−0.01 to 0.08)
Medical literacy × YNC−0.01 (−0.07 to 0.03)−0.01 (−0.07 to 0.03)
Online trust × GC0.06d (0.01 to 0.11)0.06 (0.01 to 0.11)
Online trust × GNC0.02 (−0.02 to 0.07)0.02 (−0.02 to 0.07)
Online trust × YC−0.05 (−0.10 to 0)−0.05 (−0.10 to −0.01)
Online trust × YNC0.07d (0.02 to 0.12)0.07d (0.02 to 0.12)
Trust in government × GC0.19b (0.14 to 0.24)0.20b (0.14 to 0.25)
Trust in government × GNC0.01 (−0.04 to 0.05)0.01 (−0.04 to 0.05)
Trust in government × YC0.03 (−0.02 to 0.08)0.02 (−0.02 to 0.08)
Trust in government × YNC−0.14b (−0.20 to −0.09)−0.15b (−0.20 to −0.10)

aAuthors’ analysis of revealed choices in the choice-based conjoint analysis. Health status, medical literacy, online trust, and trust in government are composite indexes, centered around mean 0 with standard deviation of 1; 95% CI are reported in parentheses.

bP<.001.

cP<.01.

dP<.05.

As shown in the last (full model) column of Table 3, the relative log odds of choosing a physician increased by 1.31 (95% CI 1.26-1.37; P<.001) and 1.32 (95% CI 1.27-1.39; P<.001) units when the government clinical ratings and commercial nonclinical ratings moved from 2 to 4 stars, respectively. The importance of these 2 types of ratings was statistically equivalent (P=.49). By comparison, the relative log odds of choosing a physician increased by a modest 1.12 (95% CI 1.07-1.18; P<.001) units when the commercial clinical ratings moved from 2 to 4 stars. The relative log odds of selecting a physician with 4 stars in nonclinical ratings provided by the government was 1.03 (95% CI 0.98-1.09; P<.001) units higher than a physician with 2 stars in this rating. The difference between the effects of government nonclinical ratings and commercial clinical ratings on patients’ choice of a primary care physician were statically significant (P=.04). The difference between the effects of clinical ratings provided by government and those provided by a commercial agency was statistically significant (P<.001). Likewise, the difference between the government clinical ratings and the government nonclinical ratings was also statistically significant (P<.001). One standard deviation improvement in a patient’s health status increased the relative log odds of choosing a physician with 4 stars in commercial nonclinical ratings by 0.18 (95% CI 0.13-0.24; P<.001) units and decreased the relative log odds of choosing a physician with 4 stars in government clinical ratings by 0.14 (95% CI 0.08-0.19; P<.001) units. Medical literacy had no statistically significant effect on how patients evaluated different types of ratings. As the level of trust in overall Web-based ratings increased, the importance of nonclinical ratings provided by a commercial agency also increased. One standard deviation increase in a patient’s trust in Web-based reviews increased the relative log odds of choosing a physician with 4 stars in nonclinical commercial ratings by 0.07 (95% CI 0.02-0.13; P=.05) units. Unsurprisingly, as the patients’ level of trust in the government increased, the importance of clinical ratings provided by government increased, whereas the importance of nonclinical ratings provided by a commercial agency decreased. One standard deviation increase in a patient’s trust in government increased the relative log odds of choosing a physician with 4 stars in government clinical ratings by 0.20 (95% CI 0.15-0.25; P<.001) units and decreased the relative log odds of choosing a physician with 4 stars in commercial nonclinical ratings by −0.15 (95% CI 0.10-0.21; P<.001) units. These trends remained consistent even when we included more variables in our model. We also examined how patients’ demographic characteristics of gender, race, income, education, marital status, and age affected the importance of each of the 4 ratings in their evaluation of primary care physicians. Table 4 presents the results. The log odds of selecting a physician with 4 stars in nonclinical government ratings relative to a physician with 2 stars was 0.23 (95% CI 0.13-0.33; P<.001) units higher for females compared with males. Similar star increase in nonclinical commercial ratings increased the relative log odds of selecting the physician by female patients by an additional 0.15 (95% CI 0.04-0.25; P=.006) units, compared with males.
Table 4

Interaction of ratings and patient characteristics. GC: clinical ratings provided by a public agency (government). GNC: nonclinical ratings provided by a public agency (government). YC: for clinical ratings provided by a commercial agency (commercial). YNC: the nonclinical ratings provided by a commercial agency (commercial).

ParameterParameter estimate (SE)
FemaleWhiteHigh incomeHigh educationMarriedAgeFull model
GC1.31a (0.05)1.24a (0.08)1.25a (0.05)1.18a (0.05)1.26a (0.05)1.15a (0.12)1.03a (0.14)
GNC0.87a (0.05)1.16a (0.08)0.98a (0.04)1.03a (0.05)0.97a (0.04)1.22a (0.11)1.25a (0.14)
YC1.15a (0.05)1.15a (0.08)1.08a (0.05)1.11a (0.05)1.07a (0.04)1.32a (0.11)1.42a (0.14)
YNC1.19a (0.05)1.30a (0.08)1.24a (0.05)1.29a (0.05)1.18a (0.05)1.21a (0.12)1.13a (0.14)
GC × Female−0.03 (0.05)−0.04 (0.05)
GNC × Female0.23a (0.05)0.23a (0.05)
YC × Female−0.09 (0.05)−0.10 (0.05)
YNC × Female0.18b (0.05)0.15b (0.05)
GC × White0.05 (0.09)0.03 (0.10)
GNC × White−0.19c (0.09)−0.17 (0.09)
YC × White−0.08 (0.09)−0.05 (0.09)
YNC × White−0.01 (0.09)−0.04 (0.10)
GC × Income0.074 (0.07)0.01 (0.08)
GNC × Income0.02 (0.07)0.02 (0.07)
YC × Income0.01 (0.07)0 (0.07)
YNC × Income0.10 (0.07)0.03 (0.08)
GC × Education0.21b (0.07)0.21b (0.07)
GNC × Education−0.06 (0.07)−0.07 (0.07)
YC × Education−0.04 (0.07)−0.04 (0.07)
YNC × Education0 (0.07)−0.01 (0.07)
GC × Married0.05 (0.07)0.03 (0.08)
GNC × Married0.05 (0.07)0.046 (0.07)
YC × Married0.04 (0.07)0.09 (0.08)
YNC × Married0.22b (0.07)0.18c (0.08)
GC × Age0.003 (0.003)0.01 (0.01)
GNC × Age−0.006c (0.002)−0.006c (0.003)
YC × Age−0.006c (0.002)−0.006c (0.003)
YNC × Age0.002 (0.003)0 (0.003)

aP<.001.

bP<.01.

cP<.05.

The relative importance of different types and sources of ratings on patients’ choice. GC: clinical ratings provided by a public agency (government). GNC: nonclinical ratings provided by a public agency (government). YC: clinical ratings provided by a commercial agency (commercial). YNC: nonclinical ratings provided by a commercial agency (commercial). aAuthors’ analysis of revealed choices in the choice-based conjoint analysis. Health status, medical literacy, online trust, and trust in government are composite indexes, centered around mean 0 with standard deviation of 1; 95% CI are reported in parentheses. bP<.001. cP<.01. dP<.05. Interaction of ratings and patient characteristics. GC: clinical ratings provided by a public agency (government). GNC: nonclinical ratings provided by a public agency (government). YC: for clinical ratings provided by a commercial agency (commercial). YNC: the nonclinical ratings provided by a commercial agency (commercial). aP<.001. bP<.01. cP<.05.

Discussion

Principal Findings

To the best of our knowledge, this was the first research that, using a conjoint analysis, uncovered how individuals used Web-based ratings to compare and choose medical providers. We found that the clinical ratings provided by the government and the nonclinical ratings provided by a commercial agency were significantly more important for patient choice than nonclinical ratings provided by the government or clinical ratings provided by commercial agencies. We also found some differences in the importance of ratings based on the sociodemographic and health characteristics of respondents. Healthier patients paid more attention to nonclinical ratings, especially those from a commercial agency. On the other hand, for healthier patients, the importance of clinical ratings, notably those that are provided by the government, was lower. We found that female patients gave more importance to nonclinical ratings provided by both public and commercial agencies, compared with males. In comparison with other races, white respondents paid less attention to the nonclinical ratings provided by government. There was no other difference between racial groups in the importance of different types of ratings in the physician choice decision. Income did not play a role in the way respondents used the ratings in their decision. As patients get older, nonclinical ratings provided by the government and the clinical ratings provided by a commercial agency became even less important in how they evaluated medical providers. A particular strength of this study was that we utilized a carefully controlled experimental design to observe the revealed preferences of participants rather than merely asking them to state them in response to a questioner, which could otherwise be subject to attribution or social desirability biases. Revealed preferences elicited in this experiment provided a more natural context, even when presented in hypothetical settings, and gave us greater confidence that the effects we observed within the sample were driven by the conjoint attributes rather than other unobserved factors.

Limitations

One limitation of our study was that we rated the attributes of the physicians by either 2 or 4 stars, whereas in reality, the ratings usually have 5 levels, between 1 and 5 stars. We limited the ratings to only 2 levels to reduce the number of possible combinations. If we considered 5 levels for each rating, the number of possible physician profiles would have surged from 16 to 625. Respondents could not reasonably compare these many physician profiles with each other. A second limitation of this study was that, in comparison with the US general population, its sample was drawn from younger, more educated, and less affluent individuals. Although samples from AMT have been shown to respond similarly to representative samples of the US population [37], the results from the study must be interpreted in light of the characteristics of the sample. Third, this study only focused on American respondents, and therefore, findings may not generalize to individuals outside of the United States. This was due to the fact that constructs such as medical literacy, health status, and trust in government significantly vary across individuals from different countries. Moreover, the presence of commercial websites and the availability of alternative government websites also vary across countries, which represents a further limitation on generalizability. Finally, in our study, we did not ask respondents whether they were familiar with the sources of information they were being asked to evaluate, primarily because our major focus was on the source (ie, government vs commercial) rather than a specific website. Future experiments could also ask respondents about their familiarity with the sources of information that they are asked to evaluate in the experiment.

Future Research

There are 3 potential areas for further research. The first is to examine how familiar individuals are with the sources of information provided by governmental and commercial agencies. Although most individuals are now fairly familiar with the commercial rating websites, knowledge about the other sources of information provided through governmental websites may be limited. It would be useful to quantify the level of awareness of such information as a precursor to designing appropriate policies to inform the public. The second is to replicate this study on an international sample to investigate how individuals outside of the United States rely on different sources and types of information for choosing their primary care physicians. Finally, the relative importance of Web-based ratings in comparison with other factors such as insurance coverage, recommendations of family and friends, and proximity to patients’ residence is still unclear and could be investigated in future research.

Policy Recommendations

The findings of this research have implications for policy makers and medical providers. Although the government has expended substantial resources on clinical quality ratings, our study indicates a need to also acknowledge the importance of nonclinical measures. This is consistent with the recent CMS efforts and policy recommendations [53] to tie reimbursements to patient satisfaction. To the extent that nonclinical ratings appear to be more important for healthier patients, it clearly underscores the important role played by the “experience” of interacting with a physician for individuals whose visits to the doctor are likely to be preventive rather than curative. Primary care providers can consider ways in which the patient’s experience can be improved, such as reduced waiting time and more empathetic interactions, which will eventually be reflected in the nonclinical ratings they receive. The results of this study could also encourage a public relations campaign to increase public awareness of the reviews that are government maintained and are more clinically based. Our result on gender differences in the relative salience of nonclinical ratings further revealed the importance of improving the patient experience for providers who are focused on women’s health services. With respect to patients’ age, we found that older patients and those who trusted government more paid more attention to government-provided ratings. This is corroborated by prior literature, which documents that citizens who trust government more are also more satisfied with government websites [54]. We therefore recommend that CMS create website content and user experiences that are tailored for Medicare beneficiaries and older patients as they rely on government-provided information more than the younger patients. Our results also indicated that commercial websites can be more successful in attracting younger individuals. If CMS intends to expand its audience, it should consider information dissemination strategies that appeal to patients in this segment. Given the recent apprehensions expressed about the quality and representativeness of ratings provided by commercial websites [55], it is a matter of some concern that patients gave equal importance to commercial ratings of nonclinical aspects of care much as they did to government ratings of clinical aspects of care. This is likely a result of the richness of the information that patients believe they can receive from other patients who have engaged in interactions with the medical provider. It might also be driven by other factors such as the first mover advantage of commercial organizations as they have been active in rating a wide variety of services earlier than other governmental agencies. To that end, our findings suggest that patients have developed a preference for commercial websites for experience-based ratings of medical providers, that is, ratings that primarily capture information about the patient’s experience with the medical provider. Thus, government agencies that offer similar ratings should pay careful attention to improve the usability of the information while concurrently addressing any perceptual obstacles that may prevent consumers from using these ratings.

Conclusions

Our research shows that patients pay equal attention to both clinical and nonclinical ratings when choosing a primary care physician. To obtain information about clinical ratings, they rely more on government sources, whereas for information on nonclinical ratings, they rely more on commercial sources. Both public and private agencies expend significant resources to design metrics, collect data, calculate ratings, and report them to the public. These resources are limited and should be optimally allocated to the type of ratings that consumers appreciate and will use the most. The findings of this research highlight the importance of efforts from government agencies such as CMS to improve its reporting of nonclinical ratings. Given the importance of nonclinical ratings in patients’ decision making, we recommend that medical providers pay close attention to their nonclinical ratings on commercial websites as they represent a consequential source of customer feedback for improving the patient experience. Ultimately, the overarching objective of all rating sources must be focused on protecting patients from incorrect or misleading data, while simultaneously educating them on how best to interpret and make best use of the information presented.
  36 in total

1.  Public awareness, perception, and use of online physician rating sites.

Authors:  David A Hanauer; Kai Zheng; Dianne C Singer; Achamyeleh Gebremariam; Matthew M Davis
Journal:  JAMA       Date:  2014-02-19       Impact factor: 56.272

2.  National hospital ratings systems share few common scores and may generate confusion instead of clarity.

Authors:  J Matthew Austin; Ashish K Jha; Patrick S Romano; Sara J Singer; Timothy J Vogus; Robert M Wachter; Peter J Pronovost
Journal:  Health Aff (Millwood)       Date:  2015-03       Impact factor: 6.301

3.  Engaged patients will need comparative physician-level quality data and information about their out-of-pocket costs.

Authors:  Jill Mathews Yegian; Pam Dardess; Maribeth Shannon; Kristin L Carman
Journal:  Health Aff (Millwood)       Date:  2013-02       Impact factor: 6.301

4.  Strategies for reporting health plan performance information to consumers: evidence from controlled studies.

Authors:  Judith H Hibbard; Paul Slovic; Ellen Peters; Melissa L Finucane
Journal:  Health Serv Res       Date:  2002-04       Impact factor: 3.402

5.  Website Characteristics and Physician Reviews on Commercial Physician-Rating Websites.

Authors:  Tara Lagu; Katherine Metayer; Michael Moran; Leidy Ortiz; Aruna Priya; Sarah L Goff; Peter K Lindenauer
Journal:  JAMA       Date:  2017-02-21       Impact factor: 56.272

6.  Are Facebook user ratings associated with hospital cost, quality and patient satisfaction? A cross-sectional analysis of hospitals in New York State.

Authors:  Lauren Campbell; Yue Li
Journal:  BMJ Qual Saf       Date:  2017-06-09       Impact factor: 7.035

7.  Understanding The Role Played By Medicare's Patient Experience Points System In Hospital Reimbursement.

Authors:  Marc N Elliott; Megan K Beckett; William G Lehrman; Paul Cleary; Christopher W Cohea; Laura A Giordano; Elizabeth H Goldstein; Cheryl L Damberg
Journal:  Health Aff (Millwood)       Date:  2016-09-01       Impact factor: 6.301

8.  Trends of online ratings of otolaryngologists: what do your patients really think of you?

Authors:  Lindsay Sobin; Parul Goyal
Journal:  JAMA Otolaryngol Head Neck Surg       Date:  2014-07       Impact factor: 6.223

9.  Consumers' Interest In Provider Ratings Grows, And Improved Report Cards And Other Steps Could Accelerate Their Use.

Authors:  Steven D Findlay
Journal:  Health Aff (Millwood)       Date:  2016-04       Impact factor: 6.301

10.  Changes in Consumer Demand Following Public Reporting of Summary Quality Ratings: An Evaluation in Nursing Homes.

Authors:  Rachel M Werner; R Tamara Konetzka; Daniel Polsky
Journal:  Health Serv Res       Date:  2016-02-11       Impact factor: 3.734

View more
  9 in total

1.  The Association Between Physician Race/Ethnicity and Patient Satisfaction: an Exploration in Direct to Consumer Telemedicine.

Authors:  Kathryn A Martinez; Kaitlin Keenan; Radhika Rastogi; Joud Roufael; Adrianne Fletcher; Mark N Rood; Michael B Rothberg
Journal:  J Gen Intern Med       Date:  2020-07-06       Impact factor: 5.128

2.  [Patient orientation of websites of otorhinolaryngologists in private practice in Bavaria].

Authors:  A Brückner; J Klewer; C Zölsmann
Journal:  HNO       Date:  2020-03       Impact factor: 1.284

3.  What Do Patients Say About Doctors Online? A Systematic Review of Studies on Patient Online Reviews.

Authors:  Y Alicia Hong; Chen Liang; Tiffany A Radcliff; Lisa T Wigfall; Richard L Street
Journal:  J Med Internet Res       Date:  2019-04-08       Impact factor: 5.428

4.  Reporting of Patient Experience Data on Health Systems' Websites and Commercial Physician-Rating Websites: Mixed-Methods Analysis.

Authors:  Tara Lagu; Caroline M Norton; Lindsey M Russo; Aruna Priya; Sarah L Goff; Peter K Lindenauer
Journal:  J Med Internet Res       Date:  2019-03-27       Impact factor: 5.428

5.  Determinants of physician networks: an ethnographic study examining the processes that inform patterns of collaboration and referral decision-making among physicians.

Authors:  Patrick Kierkegaard; Jason Owen-Smith
Journal:  BMJ Open       Date:  2021-01-05       Impact factor: 2.692

6.  Which surgeon demographic factors influence postoperative complication rates after total knee arthroplasty at U.S. News and World Report top-ranked orthopedic hospitals?

Authors:  Adam M Gordon; Andrew R Horn; Keith B Diamond; Mitchell K Ng; Matthew L Magruder; Orry Erez
Journal:  Arthroplasty       Date:  2022-07-04

7.  Identifying Influences in Patient Decision-making Processes in Online Health Communities: Data Science Approach.

Authors:  Mingda Li; Jinhe Shi; Yi Chen
Journal:  J Med Internet Res       Date:  2022-08-31       Impact factor: 7.076

8.  Correlation of Online Physician Rating Subscores and Association With Overall Satisfaction: Observational Study of 212,933 Providers.

Authors:  Hanson Hanqing Zhao; Michael Luu; Brennan Spiegel; Timothy John Daskivich
Journal:  J Med Internet Res       Date:  2020-10-27       Impact factor: 5.428

9.  Cross-sectional analysis of online patient reviews of infertility care providers.

Authors:  Ricci Allen; Shruti Agarwal; Mark P Trolice
Journal:  F S Rep       Date:  2020-07-25
  9 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.