| Literature DB >> 22648674 |
K E Hartmann1, D B Matchar, S Chang.
Abstract
Use of medical tests should be guided by research evidence about the accuracy and utility of those tests in clinical care settings. Systematic reviews of the literature about medical tests must address applicability to real-world decision-making. Challenges for reviews include: (1) lack of clarity in key questions about the intended applicability of the review, (2) numerous studies in many populations and settings, (3) publications that provide too little information to assess applicability, (4) secular trends in prevalence and the spectrum of the condition for which the test is done, and (5) changes in the technology of the test itself. We describe principles for crafting reviews that meet these challenges and capture the key elements from the literature necessary to understand applicability.Entities:
Mesh:
Year: 2012 PMID: 22648674 PMCID: PMC3364357 DOI: 10.1007/s11606-011-1961-9
Source DB: PubMed Journal: J Gen Intern Med ISSN: 0884-8734 Impact factor: 5.128
Using the PICOTS Framework to Assess and Describe Applicability of Medical Tests*
| PICOTS element | Potential characteristics to describe and assess | Challenges when assessing studies | Example | Potential systematic approaches for decisions |
|---|---|---|---|---|
| Population | ▪ Justification for lumping or splitting key questions | ▪ Source of population not described | Education/literacy level not reported in study of pencil-and-paper functional status assessment | Exclude |
| ▪ Method of identification/selection | ▪ Study population poorly specified | – Flag missing elements in tables/text | ||
| ▪ Inclusion & exclusion criteria o for the review | ▪ Key characteristics not reported | – Organize data within key questions by presence/absence of key elements | ||
| ▪ Demographic characteristics of those included in review | ▪ Unclear whether test performance varies by population | – Include presence/absence as parameter in meta-regression or sensitivity analyses | ||
| ▪ Prevalence of condition in practice and in studies | – Note need for challenge to be addressed in future research | |||
| ▪ Spectrum of disease in practice and in studies | ||||
| Intervention | ▪ Version of test used in practice and in studies | Version/ instrumentation not specified | Ultrasound machines and training of sonographers not described in study of fetal nuchal translucency assessment for detection of aneuploidy | Exclude a priori if version critical and not assessed Or include but: |
| ▪ How and by whom tests are conducted in practice and in studies | Training/quality control not described | – Contact authors for clarification | ||
| ▪ Cutoff/diagnostic thresholds applied in practice and in studies | Screening and diagnostic uses mixed | – Flag version of test or deficits in reporting in tables/text | ||
| ▪ Skill of assesors when interpretation of test required in studies | – Discuss implications | |||
| – Model cutoffs and conduct sensitivity analyses | ||||
| Comparator | ▪ Use of gold standard vs. “alloy” standard in studies | Gold standard not applied | Cardiac CT compared with stress treadmill without use of angiography as a gold standard | Exclude a priori if no gold standard Or include but: |
| ▪ Alternate or “usual” test used in the studies | Correlational data only | – Restrict to specified comparators | ||
| ▪ How test is used as part of management strategy (e.g., triage, replacement, or add-on) in practice and in studies | – Group by comparator in tables/text | |||
| ▪ In trials is comparator no testing vs. usual care with ad hoc testing | ||||
| Outcome of use of the test | ▪ How accuracy outcomes selected for review relate to use in practice: | Failure to test “normals,” or subset, with gold standard | P-value provided for mean of continuous test results by disease status but confidence bounds not provided for performance characteristics | Exclude a priori if test results cannot be mapped to disease status (i.e., 2 × 2 or other test performance data cannot be extracted) Exclude if subset of “normals” not tested Or include but: |
| ▪ Accuracy of disease status classification | Precision of estimates not provided | – Flag deficits in tables/text | ||
| • Sensitivity/specificity | Tests used as part of management strategy in which exact diagnosis is less important than “ruling out” a disease | – Discuss implications | ||
| • Predictive values | – Assess heterogeneity in meta-analysis and comment of sources of heterogeneity in estimates | |||
| • Likelihood ratios | ||||
| • Diagnostic odds ratio | ||||
| • Area under curve | ||||
| • Discriminant capacity | ||||
| Clinical Outcomes from test results | ▪ How studies addressed clinical outcomes selected for the review: | ▪ Populations and study designs of included studies heterogeneous with varied findings | Bone density testing reported in relation to fracture risk reduction without consideration of prior fracture or adjustment for age | Exclude if no disease outcomes and outcomes key to understanding intended use case Or include and: |
| • Earlier diagnosis | ▪ Data not stratified or adjusted for key predictors | – Document details of deficits in tables/text | ||
| • Earlier intervention | – Discuss implications | |||
| • Change in treatment given | – Note need for challenge to be addressed in future research | |||
| • Change in sequence of other testing | ||||
| • Change in sequence/intensity of care | ||||
| • Improved outcomes, quality of life, costs, etc. | ||||
| Timing | ▪ Timing of availability of results to care team in studies and how this might relate to practice | ▪ Sequence of use of other diagnostics unclear | D-dimer studies in which it is unclear when results were available relative to DVT imaging studies | Exclude if timing/sequence is key to understanding intended use case Or include and: |
| ▪ Placement of test in the sequence of care (e.g., relationship of test to treatment or follow-on management strategies) of studies and how this might relate to practice | ▪ Time from results to treatment not reported | – Contact authors for information | ||
| ▪ Timing of assessment of disease status and outcomes in studies | ▪ Order of testing varies across subjects and was not randomly assigned | – Flag deficits in tables/text | ||
| – Discuss implications | ||||
| – Note need for challenge to be addressed in future research | ||||
| Setting | ▪ How setting of test in studies relate to key questions and current practice: | ▪ Resources available to providers for diagnosis and treatment of condition vary widely | Diagnostic evaluation provided by geriatricians in some studies and unspecified primary care providers in others | Exclude if care setting known to influence test/outcomes or if setting is key to understanding intended use case Or include but: |
| • Primary care vs. specialty care | ▪ Provider type/specialty vary across settings | – Document details of setting | ||
| • Hospital-based | ▪ Comparability of care in international settings unclear | – Discuss implications | ||
| • Routine processing vs. specialized lab or facility | ||||
| • Specialized personnel | ||||
| • Screening vs. diagnostic use |
*Abbreviations: CT = computed tomography; DVT = deep venous thrombosis