Literature DB >> 30340488

The QUEST for quality online health information: validation of a short quantitative tool.

Julie M Robillard1,2,3, Jessica H Jun4,5, Jen-Ai Lai4,5, Tanya L Feng4,5.   

Abstract

BACKGROUND: Online health information is unregulated and can be of highly variable quality. There is currently no singular quantitative tool that has undergone a validation process, can be used for a broad range of health information, and strikes a balance between ease of use, concision and comprehensiveness. To address this gap, we developed the QUality Evaluation Scoring Tool (QUEST). Here we report on the analysis of the reliability and validity of the QUEST in assessing the quality of online health information.
METHODS: The QUEST and three existing tools designed to measure the quality of online health information were applied to two randomized samples of articles containing information about the treatment (n = 16) and prevention (n = 29) of Alzheimer disease as a sample health condition. Inter-rater reliability was assessed using a weighted Cohen's kappa (κ) for each item of the QUEST. To compare the quality scores generated by each pair of tools, convergent validity was measured using Kendall's tau (τ) ranked correlation.
RESULTS: The QUEST demonstrated high levels of inter-rater reliability for the seven quality items included in the tool (κ ranging from 0.7387 to 1.0, P < .05). The tool was also found to demonstrate high convergent validity. For both treatment- and prevention-related articles, all six pairs of tests exhibited a strong correlation between the tools (τ ranging from 0.41 to 0.65, P < .05).
CONCLUSIONS: Our findings support the QUEST as a reliable and valid tool to evaluate online articles about health. Results provide evidence that the QUEST integrates the strengths of existing tools and evaluates quality with equal efficacy using a concise, seven-item questionnaire. The QUEST can serve as a rapid, effective, and accessible method of appraising the quality of online health information for researchers and clinicians alike.

Entities:  

Keywords:  Instrument validation; Online health information; Quality evaluation; eHealth

Mesh:

Year:  2018        PMID: 30340488      PMCID: PMC6194721          DOI: 10.1186/s12911-018-0668-9

Source DB:  PubMed          Journal:  BMC Med Inform Decis Mak        ISSN: 1472-6947            Impact factor:   2.796


Background

The Internet has revolutionized how information is distributed and has led to the rapid expansion of health resources from a wide variety of content providers, ranging from government organizations to for-profit companies. Consulting online health information is an increasingly popular behavior, with 80% of Internet users engaging in this activity [1]. Health information consumers worldwide, particularly those in developing countries and remote areas, may benefit from accessible and immediate retrieval of up-to-date information [2, 3]. This new information gateway also promotes autonomy by allowing patients to be more active in their health [4]. The dynamic nature of the Internet, however, introduces important concerns in parallel with these benefits. Online information is unregulated and can be of highly variable quality [5]. This has critical implications for users as it is estimated that over half of the adult population in the United States and Canada does not possess an adequate level of health literacy [6, 7], and low health literacy is negatively correlated with the ability to discriminate between high and low quality eHealth information [8]. Compounding this issue, there is a growing number of individuals who use online information to guide health care decisions, either for themselves or on behalf of another person. It is therefore crucial to develop effective methods to evaluate online health information [9]. To this end, there have been many efforts to develop tools that assess the quality of online health information; while such tools will not solve the issue of regulation, they can assist end-users, health care professionals and researchers in differentiating between high- and low-quality online sources. A scoping review of the literature on the evaluation of health information was conducted using Arksey and O’Malley’s six-stage methodological framework [10]. The scoping review aimed to identify existing health information evaluation tools and information available in the literature on their demonstrated validity and reliability. An iterative team approach was used to determine a search strategy balancing feasibility and comprehensiveness. Data was collected via keyword searches and citation searches on Google Scholar and PubMed. Seven combinations of following keywords were used: online, health information, evaluate, evaluation, tool, quality, validity, testing, validation, and assessment. A total of 49 records were retrieved between January 15, 2016 and February 5, 2016. Thirty-six1 of these articles were included in the review based on the following inclusion criteria: 1) the article is in the English language; 2) validation of an assessment tool related to quality of health information was the focus of the article. Fifteen tools2 currently available in the literature were identified in the scoping review. A follow-up search was conducted on September 10, 2018, yielding three additional tools: the Quality Index for health-related Media Reports (QIMR) [11], the “Date, Author, References, Type, and Sponsor” (DARTS) tool [12] and Index of Scientific Quality (ISQ) [13]. The tools identified range from generic assessments, intended for use across multiple domains of online health information, to assessments targeted to a specific: 1) health condition [14, 15]; 2) aspect of a condition such as treatment [12, 16]; 3) audience [17, 18]; or 4) type of media [11, 13]. As such, a disadvantage of existing tools is that they are limited in the scope of their application. Many of the existing tools identified, with some notable exceptions, are lengthy and potentially arduous to use, out-dated, or no longer available online [3]. Some tools consist of sets of criteria or checklists that do not provide a quantitative result, making it difficult to compare information from different sources. Finally, while there are many studies evaluating online health information using existing quality evaluation tools, studies assessing the validity, reliability, and efficacy of the tools themselves are lacking in the medical informatics literature. At present, there is no clear universal standard for evaluating the quality of online health information [3]. Many researchers and regulatory bodies, including the World Health Organization, have called for the establishment of such a standard [9]. Quality criteria across existing tools often overlap and thus may serve as the basis for developing a universalized set of criteria. Aslani et al. distilled a total of 34 criteria from five evaluation tools into 10 general criteria, subdivided into four categories: author, sponsors, and individual(s) responsible for the website; purpose of the website and supporting evidence; design, ease of use, privacy, and interactibility of the website; and date of update [19]. These aggregate criteria largely correspond to groupings of criteria generated in previous reviews of the literature [20, 21]. The criteria also align with the “5 C’s” of website quality (credibility, currency, content, construction, and clarity) outlined by Roberts [22]. Of the many criteria-based assessment tools that have been developed, only a fraction have been tested for inter-rater reliability and even fewer have been validated [23]. Of tools that have reported measuring inter-rater reliability, few have consistently achieved acceptable levels of agreement across all criteria [24]. Gagliardi and Jadad [25] found that only five of 51 rating instruments they evaluated provided explicit evaluation criteria and none were validated. In a more recent review of 12 instruments by Breckons et al. [23], only two tools, DISCERN and the LIDA Minervation tool, contained any measure of reliability and validity. The DISCERN tool is the only tool currently available online for which substantive validation data is publicly available. During development of the tool, a questionnaire administered to information providers and self-help organizations was used to establish face and content validity and inter-rater reliability [16]. Additionally, external assessments indicated significant correlation with content coverage and correctness [26], good internal consistency, and significant inter-rater reliability [27]. Past comparisons to other tools, including the Mitretek Information Quality Tool (IQT) [27], Sandvik quality scale [28], EQIP [17], and DARTS [26], found significant convergent validity with DISCERN. However, DISCERN is limited in its scope of application as it is focused on treatment information and as such is not applicable to online content about other aspects of health and illness including prevention and diagnosis. There is currently no singular quantitative tool that has undergone a validation process, can be used for a broad range of health information, and strikes a balance between ease of use, concision and comprehensiveness (Fig. 1). To address these gaps, we developed the QUality Evaluation Scoring Tool (QUEST). The QUEST quantitatively measures six aspects of the quality of online health information: authorship, attribution, conflict of interest, currency, complementarity, and tone (Fig. 2), yielding an overall quality score between 0 and 28. Attribution is measured through two items, yielding a seven-item evaluation for six measures of health information quality. The criteria were chosen based on a review of existing tools used to evaluate the quality of online information by Chumber et al. [29], Sandvik et al. [28], and Silberg et al. [30]; content analysis was used to capture the overarching categories assessed by these tools [31].
Fig. 1

Review of existing quality evaluation tools (n = 16). Adapted from the CONSORT 2010 Flow Diagram available at http://www.consort-statement.org/consort-statement/flow-diagram

Fig. 2

Description of the QUEST criteria. Scores in the individual sections are weighted and summed to generate a total score of up to 28

Review of existing quality evaluation tools (n = 16). Adapted from the CONSORT 2010 Flow Diagram available at http://www.consort-statement.org/consort-statement/flow-diagram Description of the QUEST criteria. Scores in the individual sections are weighted and summed to generate a total score of up to 28 When applying the QUEST, each of the seven quality items is assigned a weighted score. The weighting of each criterion was developed based on two factors: (i) how critical it is to the overall quality of the article, established by a preliminary analysis of a sample of websites, and (ii) consideration of the criterion’s ethical implications. One criterion, attribution, is measured through a two-step process by identifying (1) the presence of references to scientific studies and, (2) the type of studies referenced, if any (e.g., animal models, observational studies, meta-analyses, clinical trials). The second item, which assigns a ranking based on the types of studies included, is in accordance with the GRADE criteria for clinical evidence [32]. This item is scored as a support to the overall quality of the health information presented, not as a judgment of the referenced studies’ quality. The aim of the present study was to evaluate whether the QUEST reliably measures a similar concept of quality to existing tools. Here we present the results of the inter-rater reliability and convergent validity analyses.

Methods

Sample

For the purposes of this study, Alzheimer disease (AD) was used as the reference health condition as there is an abundance of online articles on this topic [33, 34], and there are established methodologies for sampling in this field [31]. Online articles containing AD treatment information were retrieved using a location-disabled search on Google.com/ncr (no country redirect) to avoid localized results. Searches were conducted on an application that prevents the collection of browsing history and cookies during the search and browsing history and cookies were cleared before each search to ensure that search results were not influenced by these factors. Forty-eight different combinations of search terms related to the treatment of AD were used. Articles were extracted from the first three pages of search results, based on analyses of aggregate data on online activity patterns indicating that most Internet users tend not to view past the third page of search results [35]. Each page of search results was comprised of nine articles, totalling 27 articles for each key word combination. Inclusion criteria for the articles were: 1) the article is in the English language; 2) no payment or login is required to access the article; 3) treatment of AD is the main focus of the article as determined by the content of the headline and lead paragraph; and 4) treatment interventions discussed in the article are not solely based on animal experiments. An automatic number generator was used to obtain random 10% samples of articles that met these inclusion criteria in this present study. In a separate sample, online articles containing information about the prevention of AD were retrieved using similar methods. To retrieve these articles, 105 combinations of search terms related to AD prevention were used. Articles were screened according to criteria 1, 2, and 3 of the inclusion criteria used for treatment articles, with the exception that criterion 3 focused on prevention rather than treatment. As with the treatment-related articles, a random 10% sample of relevant articles was used for validation. In the present study, an article is defined as the heading on a webpage and the text associated with it, excluding links, images and advertising outside of the main body of text. We selected this sampling strategy based on previous investigations of inter-rater reliability and validity of similar tools that have assessed samples of 12 to 40 websites [23, 26, 27, 36, 37].

Reliability analysis

The QUEST was applied to each sample of online articles by two independent raters (JJ and TF for the prevention sample and JJ and JL for the treatment sample). Two of the three raters were naïve to tool development. To evaluate inter-rater agreement between the two reviewers, a weighted Cohen’s kappa (κ) was calculated for each item of the tool. Agreement was interpreted according to Landis and Koch, where a κ-value of 0.0 to 0.2 indicates slight agreement, 0.21 to 0.40 indicates fair agreement, 0.41 to 0.60 indicates moderate agreement, 0.61 to 0.80 indicates substantial agreement, and 0.81 to 1.0 indicates almost perfect or perfect agreement [38]. Following initial ratings of the samples, remaining disagreements were resolved by discussion to achieve 100% agreement.

Validity analysis

Three tools were selected for comparison with the QUEST based on availability, ubiquity of use, and relatedness of quality criteria and were applied to both samples. The Health on the Net Foundation’s HONcode Code of Conduct and the DISCERN instrument [16] are two of the most widely used and cited quality evaluation tools [5]. The DISCERN instrument is a 16-item questionnaire intended specifically for evaluation of health information on treatment choices, and has been found to demonstrate good inter-rater reliability and face and content validity. The HONcode Code of Conduct is a set of eight criteria used to certify websites containing health information [5]; its creators also developed a Health Website Evaluation Tool, which was used in this analysis due to its closer similarity in purpose and format to the QUEST and other tools. General quality items developed by Sandvik comprised the final tool for comparison [28]. All three tools selected for comparison are criteria-based, can be applied by a non-expert user, and contain quality criteria that, in general, align categorically with each other and the QUEST (Table 1).
Table 1

Comparison of quality items used in the QUEST, HONcode, Sandvik, and DISCERN tools

Quality criteriaQUESTHONcodeSandvikDISCERN
AttributionXXXX
CurrencyXXXX
AuthorshipXXX
BalanceXXX
ReliabilityXX
InteractivityXX
ToneX
Conflict of interestX
ComplementarityX
Mission/targetX
AudienceX
PrivacyX
Overall ReliabilityX
OwnershipX
NavigabilityX
Quality of information on treatment choicesX
Overall RatingX
Comparison of quality items used in the QUEST, HONcode, Sandvik, and DISCERN tools The QUEST and the three tools for comparison were applied to the 10% sample of treatment-related articles and the 10% sample of prevention-related articles by one investigator. The numeric scores obtained by each tool were converted to percentage scores to facilitate comparison across tools. The distribution of quality scores generated by the QUEST was plotted as a histogram to determine whether a spectrum of quality was captured by the sample (see Fig. 1, Robillard and Feng 2017 [31]). For each tool, the articles were ranked based on their scores and rankings were compared across tools in order to measure convergence. To accomplish this, a two-tailed Kendall’s tau (τ) ranked correlation [39] was used to measure convergence at α = .05. Confidence intervals (CI) of 95% for τ were calculated using Z0.05. Six correlational tests, each comparing a unique pair of tools, were performed to compare the results of the QUEST, HONcode, Sandvik, and DISCERN tools. This process was carried out for both the samples of treatment- and prevention-related articles.

Results

A total of 496 treatment articles were retrieved, with 163 of the articles meeting criteria for inclusion in the analysis and the random 10% sample consisted of 16 articles (Additional file 1). Similarly, a sample of 308 prevention articles were collected, 296 of which met inclusion criteria and 29 articles were included in the random 10% sample (Additional file 2). These articles were analyzed using the QUEST in previous quality analysis studies of articles about the prevention of AD [31]. The scores generated by each of the tools for the treatment and prevention samples are included in additional files [see Additional files 1 and 2].

Treatment

The level of inter-rater reliability was substantial between the reviewers for Attribution (κ = 0.79), high to near perfect for authorship, currency, complementarity and tone (κ ranging from 0.86 to 0.91), and perfect for type of study and conflict of interest (Table 2).
Table 2

Weighted Cohen’s kappa, standard error and 95% CI for treatment articles (n = 16)

AuthorshipAttributionType of studyConflict of interestCurrencyComplementarityTone
Observed kappa0.910.79110.860.860.91
SE0.080.1000.240.130.130.08
95% CI0.75, 10.58, 0.991, 10.32, 10.60, 10.60, 10.75, 1
Weighted Cohen’s kappa, standard error and 95% CI for treatment articles (n = 16)

Prevention

Inter-rater reliability between the two reviewers ranged from substantial to perfect agreement for each of the seven items included in the QUEST (κ ranging from 0.74 to 1.0; Table 3).
Table 3

Weighted Cohen’s kappa, standard error and 95% CI for prevention articles (n = 29)

AuthorshipAttributionType of studyConflict of interestCurrencyComplementarityTone
Observed kappa0.880.890.890.7410.750.95
SE0.090.060.060.1600.140.04
95% CI0.71, 10.78, 10.77, 10.43, 11, 10.49, 10.86, 1
Weighted Cohen’s kappa, standard error and 95% CI for prevention articles (n = 29) Scores obtained using HONcode had the widest range, 15–100%. Scores obtained using the Sandvik criteria had a narrower range, 43–100%. The DISCERN instrument returned the narrowest range of scores, 45–86%. The QUEST generated a range of scores (25–100%) wider than those generated by both the DISCERN tool and Sandvik criteria, but narrower than that of HONcode. The median percentage scores returned by the DISCERN and HONcode tools were 59% and 62% respectively, while the Sandvik criteria generated a median score of 86%. Again, the median score generated by the QUEST, 71%, fell between those of the other instruments. Quality analysis of the prevention-related articles generated similar results. HONcode generated the widest range of scores (22–100%), while DISCERN returned the narrowest range (30–88%). The range of scores obtained using the Sandvik criteria (29–93%) fell between the ranges generated by the HONcode and DISCERN instruments. The QUEST generated a range of scores (29–96%) wider than those of DISCERN and Sandvik, but narrower than that of HONcode. On the lower end, the median percentage score obtained using the DISCERN criteria was 54%. On the upper end, the median score generated by HONcode was 68%. Between these values, both the Sandvik criteria and the QUEST returned a median score of 64%. Of the six correlational tests performed between unique pairs of tools on the articles related to treatment, all six tests demonstrated a significant correlation between the tools (Table 4). Values of τ ranged from 0.47 (QUEST and HONcode) and 0.53 (HONcode and Sandvik) on the lower end to 0.62 (QUEST and Sandvik) and 0.65 (QUEST and DISCERN) on the higher end (P < .05 for all tests).
Table 4

Kendall’s tau, standard error, 95% CI, and P-value of each test for treatment articles (n = 16)

Kendall’s tau (95% CI)SEP-value
QUEST vs HONcode0.47 (0.09–0.85)0.190.015
QUEST vs Sandvik0.62 (0.23–1.01)0.200.002
QUEST vs DISCERN0.65 (0.28–1.02)0.19< 0.001
HONcode vs Sandvik0.53 (0.13–0.92)0.200.009
HONcode vs DISCERN0.58 (0.20–0.96)0.190.003
Sandvik vs DISCERN0.58 (0.19–0.96)0.200.004
Kendall’s tau, standard error, 95% CI, and P-value of each test for treatment articles (n = 16) Similarly, all six correlational tests performed on the prevention sample demonstrated a significant correlation between the tools (P < .05; Table 5). The weakest correlations were found between Sandvik and DISCERN, and the QUEST and DISCERN, which produced τ- values of 0.41 and 0.55 respectively. The strongest correlations were found between the QUEST and Sandvik (τ =0.62) and the QUEST and HONcode (τ = 0.64).
Table 5

Kendall’s tau, standard error, 95% CI, and P-value of each test for prevention articles (n = 29)

Kendall’s tau (95% CI)SEP-value
QUEST vs HONcode0.64 (0.37–0.99)0.14< 0.001
QUEST vs Sandvik0.62 (0.34–0.90)0.14< 0.001
QUEST vs DISCERN0.55 (0.29–0.82)0.14< 0.001
HONcode vs Sandvik0.61 (0.33–0.89)0.14< 0.001
HONcode vs DISCERN0.57 (0.31–0.84)0.14< 0.001
Sandvik vs DISCERN0.41 (0.13–0.68)0.140.004
Kendall’s tau, standard error, 95% CI, and P-value of each test for prevention articles (n = 29)

Discussion

In the present study to validate a novel tool to assess the quality of health information available on the Internet, we find the QUEST to have high inter-rater reliability and convergent validity when applied to two samples of online articles on AD. The results of the validity analysis of treatment and prevention samples indicate that the rankings of quality scores generated by the QUEST converge with those generated by three other tools – the HONcode Health Website Evaluation tool, the DISCERN instrument, and the Sandvik criteria. For the sample of articles on AD treatment, the strong correlation between the QUEST and the DISCERN instrument suggests that these tools evaluate a similar concept of quality. As past findings indicate that the DISCERN tool is itself a valid tool for assessing treatment information, its high level of convergence with the QUEST confers promising preliminary evidence for the validity of the QUEST. One limitation of the DISCERN tool is the ambiguity in applying a Likert scale to the data. The QUEST addresses this limitation by providing specific descriptions of the criteria for each possible score for a given item. The QUEST’s lower level of convergence with the HONcode’s evaluation of treatment-related articles may indicate a wider gap between interpretations of the concept of quality evaluated by these two tools. The HONcode tool places emphasis on aspects that are not assessed by the QUEST, such as the website’s mission, target audience, privacy policy, and interactivity [40], all of which expand on the concept of quality but increase the time required to apply the tool. However, there may be other factors that account for the discrepancy between the tools’ rankings. There exist some ambiguities in scoring websites using HONcode that are intrinsic to the design of the tool. For example, with a few exceptions, the HONcode rates questions on a dichotomous scale (Yes/No). This rating system, unlike the Likert-type scales used by the QUEST, DISCERN, and Sandvik [28], does not allow for an assessment beyond an absence or presence of criteria. Finally, some criteria are only marginally or not applicable to many websites’ content. For example, one question asks the responder to evaluate banner content, and website design has moved away from these types of site elements. Analysis of the scores generated from the sample of prevention-related articles found the strongest correlation between the QUEST and HONcode. Conversely, the QUEST displayed the poorest convergence with the DISCERN instrument. The discrepancy between these findings and those from the treatment sample, which found the strongest convergence between the QUEST and DISCERN and the weakest between the QUEST and HONcode, may reflect intrinsic differences in the purpose of the tools. The DISCERN instrument was developed specifically for the quality evaluation of treatment information, whereas the QUEST, HONcode and Sandvik criteria were developed for health information more broadly. Overall findings demonstrate a high degree of inter-rater reliability for all seven items of the QUEST. In their evaluation of the DISCERN instrument, Charnock et al. [16] found that lower agreement scores were generally associated with criteria that required more subjective assessment, such as ratings about areas of uncertainty or questions requiring scaled responses. Results from the current study indicate that more subjective items in the QUEST, such as attribution, conflict of interest and tone, achieve about equal or higher levels of inter-reliability as more objective items. Results from the reliability analysis suggest that the QUEST criteria may serve as an effective framework for current as well as future iterations of quality evaluation resources. The QUEST offers three main advantages over existing tools. Foremost, the QUEST condenses a wide range of quality evaluation criteria into a brief, seven-item questionnaire that evaluates quality with comparable efficacy to established tools. This concise design in conjunction with a weighted criteria approach facilitates the rapid evaluation of health information for a diverse group of users. For example, health care professionals may use the QUEST to evaluate the quality of information brought to them by their patients or to find high-quality articles to recommend. The QUEST may also be of value to the scientific community as it can be used as a research tool to quickly and accurately evaluate quality, facilitating the characterization and comparison of large amounts of information. Additionally, the QUEST may help inform creators of online health content, including government, industry, university, and advocacy groups, during the content development process. In terms of content, the QUEST tool is differentiated from the three other tools included in the present analysis in its weighted measurement of tone, conflict of interest, and complementarity (see Table 1). These criteria address factors such as potential bias linked to promotion of a product or intervention, whether support of the patient-physician relationship is referenced, and whether the information is presented in a balanced way. Finally, the QUEST was designed for application to a variety of health topics including information on both treatment and prevention, as well as general health information. Altogether these characteristics, combined with evidence of the QUEST’s reliability and validity, are reflective of a versatile tool suited to meet diverse user needs. It is important to note that each individual item provides information about only a single aspect of information quality, and thus the QUEST should be used as a gestalt to provide an overall assessment of quality. It should be noted that while the QUEST is designed to be a concise and universally applicable tool, there is a range of other evaluation tools in the literature with different and potentially complementary aims to QUEST (please see Appendix 2 for a comparison of currently available tools to QUEST). For example, the QIMR tool released in 2017 may be more suited for evaluating health research reports in the lay media and the AGREE instrument may be best suited to evaluating the quality of clinical practice guidelines. While the versatility of the QUEST tool lies in its applicability to a range of online health information, is not necessarily the only or most suitable tool for all types of health-related media. The focused area of the samples used in this study addresses an important and growing issue relating to the quality of online health information targeted toward aging populations, who face unique challenges in cognition which can be exacerbated by low health literacy [41]. Additionally, older adults tend to have less experience conducting online searches and critically evaluating the credibility of online information [42, 43]. Due to this combination of factors, this demographic of health consumers may be more susceptible to misinformation online. Beyond the focus on AD used for this validation study, the QUEST will benefit from further testing across a wider range of health conditions. The study design has several strengths. The correlational method used does not rely on an assumption of normality of the data, and the magnitudes of the correlation coefficients indicate the strength of correlation between the tools being compared [39]. We conducted more than one analysis on the data, comparing the QUEST to three well-established and well-regarded evaluation tools. Careful selection of tools for comparison and use of multiple tools in the analysis both contribute to the rigour of the study. However, we also recognize the limitations of the study. A sample of convenience of a relatively small number of articles was used, taken from existing collections of AD treatment and prevention articles. Due potentially to the small sample size of articles used, the Kendall’s tau scores have substantially overlapping confidence intervals; this indicates a need for further validation studies that include a larger number of articles on other health conditions and on types of health information beyond treatment and prevention, such as descriptions of symptoms and management. Furthermore, our study included only three raters, whereas it may be useful to include more raters in the future when assessing inter-rater reliability. It may also be informative to assess the predictive validity of real-life application of the tool. This may be used to predict whether sustained use of the instrument is associated with higher levels of user knowledge, engagement with care providers on the health topic, or self-efficacy in management of the health condition researched. Additionally, existing quality evaluation tools generally adopt the perspective of the health care professional in conceptualizing quality [27]. We recognize that the QUEST tool, currently aimed at health care professionals and researchers, falls into this category. Given the time constraints of clinical visits, health care professionals may not be able to assess the quality of online resources during the consultation. To address this issue, attempts have been made to automate tools such as the HONcode and the QUEST [44, 45]. Further, research indicates that the methods used by health consumers to search and appraise online health information differ from the systematic methods used by investigators [46]. As a partially non-academic area of research, a number of health information evaluation tools are not detailed or evaluated in the peer-reviewed literature and may have been excluded from the scoping review presented here. Existing efforts to expand the user base for quality evaluation tools include the HONcode Health Website Evaluation Tool and Provost et al.’s 95-item WebMedQual assessment [47]. This body of work can be expanded upon in the academic space by standardising and ensuring validity of the broad range of heterogeneous tools that exist outside of this space. Future work should continue efforts to develop a more accessible and concise patient-friendly tool that incorporates the values of end-users when assessing online health information, such as privacy and usability factors. To address this need, we are currently in the process of developing a public-friendly adaptation of the existing QUEST criteria that can be easily understood and applied by non-expert users. Finally, a novel tool aiming to address the issue of misinformation online – whether intended for use by expert or non-expert users – needs to be supplemented by a careful examination of the drivers of public attitudes toward key issues in health care. Studies have shown that social beliefs and attitudes related to a range of health issues (e.g., vaccination uptake [48, 49], health and wellbeing in an ageing population [50], uptake of mental health care [51, 52]) pose significant challenges in obtaining optimal public health outcomes. Tools such as QUEST are designed as downstream interventions that can aid health consumers and providers in differentiating between high- and low-quality information online. It is unlikely that the wide availability of these tools will be effective as a standalone intervention; additional work is required to contextualize the public spaces in which these evaluation tools will be useful and to determine how these tools can best be used in complement to health communication strategies and more upstream, systemic interventions in order to change health behaviours and attitudes.

Conclusions

Developed to address gaps in available quality evaluation tools for online health information, the QUEST is composed of a short set of criteria that can be used by health care professionals and researchers alike. Our findings demonstrate the QUEST’s reliability and validity in evaluating online articles on AD treatment and prevention. For example, two similar tools used for comparison, the DISCERN and HONcode Health Website Evaluation tools, are 12–16 questions in length. This study provides evidence that the QUEST builds on the strengths of existing instruments and evaluates quality with similar efficacy using a rapid seven-item questionnaire. As a result, this tool may serve as a more accessible resource that effectively consolidates the quality criteria outlined in previous work. Additionally, due to its simplicity and unique weighting approach, the QUEST reduces the need for users’ subjective judgment and indicates potential for future iterations of the tool to be easily tailored to the needs of different users. Based on the current evidence, the QUEST can be used to reliably assess online sources of information on treatment and prevention of AD. Following formal establishment of its reliability and validity across a wide range of health topics, the QUEST may serve as or inform a universal standard for the quality evaluation of online health information. Quality scores of treatment articles (n = 16). Scores generated by the QUEST, HONcode, Sandvik, and DISCERN tools for the 16 articles containing information on the treatment of AD. (XLSX 57 kb) Quality scores of prevention articles (n = 29). Scores generated by the QUEST, HONcode, Sandvik, and DISCERN tools for the 29 articles containing information on the prevention of AD. (XLSX 57 kb)
Table 6

Characteristics of articles (n = 36) retrieved between January 15, 2016 and February 5, 2016 using the following search terms on Google Scholar and PubMed: online, health information, evaluate, evaluation, tool, quality, validity, testing, validation, and assessment and meeting the following inclusion criteria: 1) the article is in the English language; 2) validation of an assessment tool related to quality of health information was the focus of the article

Focus of articleNumber of articlesArticle titleAuthor(s)Date of Publication
Observational or descriptive paper5Assessing, controlling, and assuring the quality of medical information on the internet: Caveant lector et viewor—let the reader and viewer bewareSilberg WM, Lundberg GD, and Musacchio RA1997
The Health On the Net Code of Conduct for Medical and Health WebsitesBoyer, C., M. Selby, J. R. Scherrer, and R. D. Appel1998
Emerging Challenges in Using Health Information from the InternetTheodosiou, Louise, and Jonathan Green2003
Health information and the internet: The 5 Cs website evaluation toolRoberts, Lorraine2010
Quality of patient health information on the Internet: reviewing a complex and evolving landscapeFahy, Eamonn, Rohan Hardikar, Adrian Fox, and Sean Mackay2014
Evaluation of quality of information using tool(s)7Health information and interaction on the internet: a survey of female urinary incontinenceSandvik, Hogne1999
Evaluation of Websites that Provide Information on Alzheimer’s DiseaseBouchier, H., and P. A. Bath2003
Accuracy of internet recommendations for prehospital care of venomous snake bitesBarker et al2010
The quality of online antidepressant drug information: An evaluation of English and Finnish language Web sitesPrusti, Marjo, Susanna Lehtineva, Marika Pohjanoksa-Mäntylä, and J. Simon Bell2012
Evaluation of dengue-related health information on the InternetRao et al2012
A Methodology to Analyze the Quality of Health Information on the Internet The Example of Diabetic NeuropathyChumber, Sundeep, Jörg Huber, and Pietro Ghezzi2014
Evaluation of Online Health Information on Clubfoot Using the DISCERN ToolKumar, Venkatesan S., Suresh Subramani, Senthil Veerapan, and Shah A. Khan2014
Development of tool12DISCERN: an instrument for judging the quality of written consumer health information on treatment choices.Charnock, D., S. Shepperd, G. Needham, and R. Gann1999
Development of a self-assessment method for patients to evaluate health information on the InternetJones J.1999
Development and Application of a Tool Designed to Evaluate Web Sites Providing Information on Alzheimer’s DiseaseBath, P. A., and H. Bouchier2003
Development and validation of an international appraisal instrument for assessing the quality of clinical practice guidelines: the AGREE projectCluzeau et al.2003
Design and testing of a tool for evaluating the quality of diabetes consumer-information web sites. Journal of Medical Internet ResearchSeidman, Joshua J, Donald Steinwachs, and Haya R Rubin2003
The development of QUADAS: a tool for the quality assessment of studies of diagnostic accuracy included in systematic reviewsWhiting, Penny, Anne WS Rutjes, Johannes B. Reitsma, Patrick MM Bossuyt, and Jos Kleijnen2003
Ensuring Quality Information for Patients: Development and Preliminary Validation of a New Instrument to Improve the Quality of Written Health Care InformationMoult, Beki, Linda S Franck, and Helen Brady2004
A model for online consumer health information qualityStvilia, Besiki, Lorri Mon, and Yong Jeong Yi2009
Health Literacy INDEX: Development, Reliability, and Validity of a New Tool for Evaluating the Health Literacy Demands of Health Information MaterialsKaphingst et al.2012
Measuring the quality of Patients’ goals and action plans: development and validation of a novel toolTeal, Cayla R., Paul Haidet, Ajay S. Balasubramanyam, Elisa Rodriguez, and Aanand D. Naik2012
The Communication AssessmenT Checklist in Health (CATCH): a tool for assessing the quality of printed educational materials for clinicians.Genova, Juliana, Isaac Nahon-Serfaty, Selma Chipenda Dansokho, Marie-Pierre Gagnon, Jean-Sébastien Renaud, and Anik M. C. Giguère2014
Development and Validation of the Guide for Effective Nutrition Interventions and Education (GENIE): A Tool for Assessing the Quality of Proposed Nutrition Education ProgramsHand, Rosa K., Jenica K. Abram, Katie Brown, Paula J. Ziegler, J. Scott Parrott, and Alison L. Steiber2015
Evaluation of tool(s)9Published Criteria for Evaluating Health Related Web Sites: ReviewKim, Paul, Thomas R. Eng, Mary Jo Deering, and Andrew Maxfield1999
Examination of instruments used to rate quality of health information on the internet: chronicle of a voyage with an unclear destinationGagliardi, Anna, and Alejandro R. Jadad2002
Evaluating the reliability and validity of three tools to assess the quality of health information on the Internet.Ademiluyi, Gbogboade, Charlotte E Rees, and Charlotte E Sheard2003
The Evaluation Criteria of Internet Health InformationKang, Nam-Mi, Sukhwa Kim, Seungkuen Hong, Seewon Ryu, Hye-Jung Chang, and Jeongeun Kim2006
Assessing the Quality of Websites Providing Information on Multiple Sclerosis: Evaluating Tools and Comparing SitesHarland, Juliet, and Peter Bath2007
What Do Evaluation Instruments Tell Us About the Quality of Complementary Medicine Information on the Internet?Breckons, Matthew, Ray Jones, Jenny Morris, and Janet Richardson2008
Tools Used to Evaluate Written Medicine and Health Information Document and User PerspectivesLuk, Alice, and Parisa Aslani2011
Tools for Assessing the Quality and Accessibility of Online Health Information: Initial Testing among Breast Cancer WebsitesWhitten, Pamela, Samantha Nazione, and Carolyn Lauckner2013
Web-site evaluation tools: a case study in reproductive health informationAslani, Azam, Omid Pournik, Ameen Abu-Hanna, and Saeid Eslami2014
Systematic literature review of tools3Empirical Studies Assessing the Quality of Health Information for Consumers on the World Wide Web: A Systematic ReviewEysenbach et al.2002
Online Health Information Tool Effectiveness for Older Patients: A Systematic Review of the LiteratureBolle, Sifra, Julia C. M. van Weert, Joost G. Daams, Eugène F. Loos, Hanneke C. J. M. de Haes, and Ellen M. A. Smets.2015
Quality of Health Information for Consumers on the Web: A Systematic Review of Indicators, Criteria, Tools, and Evaluation ResultsZhang, Yan, Yalin Sun, and Bo Xie2015
Table 7

Comparison of evaluation tools previously described in the literature and QUEST

Name of toolFocusCriteriaFormat
0QUality Evaluation Scoring Tool (QUEST)Quality of online health informationAuthorship, attribution, conflict of interest, complementarity, currency, tone6 questions rated on a scale of 0–2 or 0–1 and differentially weighted, yielding an overall quality score between 0 and 28
1DISCERNQuality of written information about treatment choicesReliability, balance, dates, source, quality of information on treatment sources, overall rating15 questions rated on a scale of 1–5
2EQIP: Ensuring Quality Information for PatientsQuality of written patient information applicable to all information typesClarity, patient-oriented design, currency, attributon, conflict of interest, completeness20 questions rated Y/Partly/N with an equation to generate a % score
3Jones’ Self-Assessment MethodSelf-assessment tool for patients to evaluate quality and relevance of health care oriented websitesContent, design, communication, and credibility9 broad questions based on 4 criteria rated Yes/No/NA
4Health on the Net Foundation’s HONcode Patient Evaluation ToolPatient evaluation tool for health-related websitesAuthorship, attribution, currency, reliability, balance, mission/target audience, privacy, interactivity, overall reliability16-item interactive questionnaire returning a % score
5Silberg standardsStandards of quality for online medical information for consumers and professionalsAuthorship, attribution, disclosure, currencySet of core standards; no score is generated
6Sandvik’s General Quality CriteriaGeneral quality measure for online health informationOwnership, authorship, source, currency, interactivity, navigability, balance7 questions rated on a scale of 0–2
7Health Information Technology Institute (HITI) Information Quality Tool *No longer availableQuality measure for health-related websitesCredibility, content, disclosure, links, design, interactivity Not available
85 C’s website evaluation toolStructured guide to systematically evaluating websites; specifically developed for nurses to use in patient care and educationCredibility, currency, content, construction, claritySeries of 36 open-ended and yes/no questions grouped under the “5 C’s”; no score is generated
9Health Literacy INDEXTool to evaluate the health literacy demands of health information materialsPlain language, clear purpose, supporting graphics, user involvement, skill-based learning, audience appropriateness, instructions, development details, evaluation methods, strength of evidence63 indicators/criteria rated yes/no, yielding criterion-specific scores and an overall % score
10Bath and Bouchier’s evaluation toolTool to evaluate websites providing information on Alzheimer’s diseaseGeneral details, information for carers, currency, ease of use, general conclusions47 questions scored from 0 to 2, generating an overall % score
11Seidman quality evaluation toolQuality of diabetes consumer-information websitesExplanation of methods, validity of methods, currency, comprehensiveness, accuracy7 structural measures and 34 performance measures, generating composite scores by section and an overall score
12Appraisal of Guidelines, REsearch and Evaluation (AGREE) Collaboration instrumentQuality of clinical practice guidelinesScope and purpose, stakeholder involvement, rigour of development, clarity and presentation, applicability, editorial independence23 items grouped into six quality domains with a 4 point Likert scale to score each item
13Communication AssessmenT Checklist in Health (CATCH) toolQuality of printed educational materials for cliniciansAppearance, layout and typography, clarity of content, language and readability, graphics, risk communication, scientific value, emotional appeal, relevance, social value/source credibility, social value/usefulness for the clinician, social value/usefulness for the health care system (hospital or government)55 items nested in 12 concepts, each rated yes/no, generating concept-specific and overall scores
14LIDA Minervation toolEvaluates the design and content of healthcare websitesAccessibility, usability (clarity, consistency, functionality, engagability), reliability (currency, conflict of interest, content production)41 questions scored on a scale of 0–3, yielding a total % score
15Mitretek Information Quality Tool (IQT) *no longer availableEvaluates information quality of online health informationAuthorship, sponsorship, currency, accuracy, confidentiality, navigability21 questions rated yes/no and weighted according to importance, generating a total score between 0 to 4
16“Date, Author, References, Type, and Sponsor” (DARTS)Assists patients in appraising the quality of online medicines informationCurrency, authorship, credibility, purpose, conflict of interestA series of six guiding questions; no score generated
17Quality Index for health-related Media Reports (QIMR)Monitors the quality of health research reports in the lay mediaBackground, sources, results, context, validity17 items rated on a 0–6 Likert scale with an 18th global rating
18Index of Scientific Quality (ISQ)Index of scientific quality for health reports in the lay pressApplicability, opinions vs. facts, validity, magnitude, precision, consistency, consequences7 items rated on a 1–5 Likert scale with an 8th global rating
  39 in total

Review 1.  Published criteria for evaluating health related web sites: review.

Authors:  P Kim; T R Eng; M J Deering; A Maxfield
Journal:  BMJ       Date:  1999-03-06

Review 2.  Empirical studies assessing the quality of health information for consumers on the world wide web: a systematic review.

Authors:  Gunther Eysenbach; John Powell; Oliver Kuss; Eun-Ryoung Sa
Journal:  JAMA       Date:  2002 May 22-29       Impact factor: 56.272

3.  The initial development of the WebMedQual scale: domain assessment of the construct of quality of health web sites.

Authors:  Mélanie Provost; Dayin Koompalum; Diane Dong; Bradley C Martin
Journal:  Int J Med Inform       Date:  2005-09-19       Impact factor: 4.046

Review 4.  Quality of patient health information on the Internet: reviewing a complex and evolving landscape.

Authors:  Eamonn Fahy; Rohan Hardikar; Adrian Fox; Sean Mackay
Journal:  Australas Med J       Date:  2014-01-31

Review 5.  Web-site evaluation tools: a case study in reproductive health information.

Authors:  Azam Aslani; Omid Pournik; Ameen Abu-Hanna; Saeid Eslami
Journal:  Stud Health Technol Inform       Date:  2014

Review 6.  Ensuring quality information for patients: development and preliminary validation of a new instrument to improve the quality of written health care information.

Authors:  Beki Moult; Linda S Franck; Helen Brady
Journal:  Health Expect       Date:  2004-06       Impact factor: 3.377

7.  Design and testing of a tool for evaluating the quality of diabetes consumer-information Web sites.

Authors:  Joshua J Seidman; Donald Steinwachs; Haya R Rubin
Journal:  J Med Internet Res       Date:  2003-11-27       Impact factor: 5.428

8.  Aging 2.0: health information about dementia on Twitter.

Authors:  Julie M Robillard; Thomas W Johnson; Craig Hennessey; B Lynn Beattie; Judy Illes
Journal:  PLoS One       Date:  2013-07-26       Impact factor: 3.240

9.  The development and validation of an instrument to measure the quality of health research reports in the lay media.

Authors:  Dena Zeraatkar; Michael Obeda; Jeffrey S Ginsberg; Jack Hirsh
Journal:  BMC Public Health       Date:  2017-04-20       Impact factor: 3.295

Review 10.  Under-vaccinated groups in Europe and their beliefs, attitudes and reasons for non-vaccination; two systematic reviews.

Authors:  N Fournet; L Mollema; W L Ruijs; I A Harmsen; F Keck; J Y Durand; M P Cunha; M Wamsiedel; R Reis; J French; E G Smit; A Kitching; J E van Steenbergen
Journal:  BMC Public Health       Date:  2018-01-30       Impact factor: 3.295

View more
  18 in total

1.  Quality evaluation of HPV vaccine-related online messages in China: a cross-sectional study.

Authors:  Wanzhou Wang; Jinliang Lyu; Mintao Li; Yunjing Zhang; Zhihu Xu; Yuanyuan Chen; Jiangjie Zhou; Shengfeng Wang
Journal:  Hum Vaccin Immunother       Date:  2020-10-15       Impact factor: 3.452

2.  Digital Scientific Platform for Independent Content in Neurology: Rigorous Quality Guideline Development and Implementation.

Authors:  Daniel Kantor; Martin Farlow; Albert Ludolph; Joan Montaner; Raman Sankar; Robert N Sawyer; Fabrizio Stocchi; Agnès Lara; Sarah Clark; Karine Deschet; Loucif Ouyahia; Yacine Hadjiat
Journal:  Interact J Med Res       Date:  2022-06-09

3.  Communication on drug safety-related matters to patients: is it even more significant in this digital era?

Authors:  Jimmy Jose
Journal:  Ther Adv Drug Saf       Date:  2020-04-09

4.  Online Information on Electronic Cigarettes: Comparative Study of Relevant Websites From Baidu and Google Search Engines.

Authors:  Ting Chen; Sarah Gentry; Dechao Qiu; Yan Deng; Caitlin Notley; Guangwen Cheng; Fujian Song
Journal:  J Med Internet Res       Date:  2020-01-24       Impact factor: 5.428

5.  Benefits of Structured Engagement with Target Audiences of a Health Website: Study Design for a Multi-Case Study.

Authors:  Jennifer Tieman; Virginia Lewis
Journal:  Healthcare (Basel)       Date:  2021-05-18

6.  Evaluating Local Multilingual Health Care Information Environments on the Internet: A Pilot Study.

Authors:  Russell Miller; Nicholas Doria-Anderson; Akira Shibanuma; Jennifer Lisa Sakamoto; Aya Yumino; Masamine Jimba
Journal:  Int J Environ Res Public Health       Date:  2021-06-25       Impact factor: 3.390

7.  Quality of web-based information at the beginning of a global pandemic: a cross-sectional infodemiology study investigating preventive measures and self care methods of the coronavirus disease 2019.

Authors:  Jenny Stern; Susanne Georgsson; Tommy Carlsson
Journal:  BMC Public Health       Date:  2021-06-14       Impact factor: 3.295

Review 8.  Review of Web-Based Toolkits for Health Care Practitioners Working With Women and Girls Affected by or at Risk of Female Genital Mutilation/Cutting.

Authors:  Hina Shaikh; Karen A McDonnell
Journal:  J Prim Care Community Health       Date:  2020 Jan-Dec

Review 9.  Tools to Assess the Trustworthiness of Evidence-Based Point-of-Care Information for Health Care Professionals: Systematic Review.

Authors:  Gerlinde Lenaerts; Geertruida E Bekkering; Martine Goossens; Leen De Coninck; Nicolas Delvaux; Sam Cordyn; Jef Adriaenssens; Patrick Vankrunkelsven
Journal:  J Med Internet Res       Date:  2020-01-17       Impact factor: 5.428

10.  Boxing for Parkinson's Disease: Has Implementation Accelerated Beyond Current Evidence?

Authors:  Meg E Morris; Terry D Ellis; Dana Jazayeri; Hazel Heng; Andrea Thomson; Arun Prasad Balasundaram; Susan C Slade
Journal:  Front Neurol       Date:  2019-12-04       Impact factor: 4.003

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.