| Literature DB >> 36186253 |
Deepak Dhamnetiya1, Ravi Prakash Jha1, Shalini Shalini2, Krittika Bhattacharyya3.
Abstract
Diagnostic tests are pivotal in modern medicine due to their applications in statistical decision-making regarding confirming or ruling out the presence of a disease in patients. In this regard, sensitivity and specificity are two most important and widely utilized components that measure the inherent validity of a diagnostic test for dichotomous outcomes against a gold standard test. Other diagnostic indices like positive predictive value, negative predictive value, positive likelihood ratio, negative likelihood ratio, accuracy of a diagnostic test, and the effect of prevalence on various diagnostic indices have also been discussed. We have tried to present the performance of a classification model at all classification thresholds by reviewing the receiver operating characteristic (ROC) curve and the depiction of the tradeoff between sensitivity and (1-specificity) across a series of cutoff points when the diagnostic test is on a continuous scale. The area under the ROC (AUROC) and comparison of AUROCs of different tests have also been discussed. Reliability of a test is defined in terms of the repeatability of the test such that the test gives consistent results when repeated more than once on the same individual or material, under the same conditions. In this article, we have presented the calculation of kappa coefficient, which is the simplest way of finding the agreement between two observers by calculating the overall percentage of agreement. When the prevalence of disease in the population is low, prospective study becomes increasingly difficult to handle through the conventional design. Hence, we chose to describe three more designs along with the conventional one and presented the sensitivity and specificity calculations for those designs. We tried to offer some guidance in choosing the best possible design among these four designs, depending on a number of factors. The ultimate aim of this article is to provide the basic conceptual framework and interpretation of various diagnostic test indices, ROC analysis, comparison of diagnostic accuracy of different tests, and the reliability of a test so that the clinicians can use it effectively. Several R packages, as mentioned in this article, can prove handy during quantitative synthesis of clinical data related to diagnostic tests. The Indian Association of Laboratory Physicians. This is an open access article published by Thieme under the terms of the Creative Commons Attribution-NonDerivative-NonCommercial License, permitting copying and reproduction so long as the original work is given appropriate credit. Contents may not be used for commercial purposes, or adapted, remixed, transformed or built upon. ( https://creativecommons.org/licenses/by-nc-nd/4.0/ ).Entities:
Keywords: ROC curve; diagnostic accuracy; diagnostic study; kappa statistics; sensitivity; specificity
Year: 2021 PMID: 36186253 PMCID: PMC9519267 DOI: 10.1055/s-0041-1734019
Source DB: PubMed Journal: J Lab Physicians ISSN: 0974-2727
Diagnostic test results in relation to true disease status in a 2 × 2 table
| Disease | ||||
|---|---|---|---|---|
| Present | Absent | Total | ||
| Screening test/New test | Positive | a (true positive) | b (false positive) | a + b |
| Negative | c (false negative) | d (true negative) | c + d | |
| Total | a + c | b + d |
| |
Diagnostic test results for the given example in relation to true disease status
| Disease | ||||
|---|---|---|---|---|
| Present | Absent | Total | ||
| Test X | Positive | 69 | 25 | 94 |
| Negative | 11 | 95 | 106 | |
| Total | 80 | 120 | 200 | |
Effect of change in prevalence on various diagnostic indices of test X
| Prevalence (%) | Sensitivity | Specificity | LR+ | LR– | PPV | NPV | Accuracy |
|---|---|---|---|---|---|---|---|
| 20 | 86.25 | 79.17 | 4.14 | 0.17 | 50.86 | 95.84 | 80.58 |
| 30 | 86.25 | 79.17 | 4.14 | 0.17 | 63.95 | 93.07 | 81.29 |
| 40 | 86.25 | 79.17 | 4.14 | 0.17 | 73.40 | 89.62 | 82.00 |
| 50 | 86.25 | 79.17 | 4.14 | 0.17 | 80.54 | 85.20 | 82.71 |
| 60 | 86.25 | 79.17 | 4.14 | 0.17 | 86.13 | 79.33 | 83.42 |
Abbreviations: LR, likelihood ratio; NPV, negative predictive value; PPV, positive predictive value.
Fig. 1Receiver operator characteristics curve of test X1 for diagnosing disease A.
Fig. 2Precision–recall curve of diagnostic test X1 for disease A.
Fig. 3Receiver operator characteristics curve of test X1, X2, and X3 for diagnosing disease A.
Summary table of various diagnostic indices and AUROC of diagnostic test X1, X2, and X3 for disease A
| Variable | Youden's index J | Sensitivity (%) | Specificity (%) | LR+ | LR– | AUC (95% CI) |
|---|---|---|---|---|---|---|
| X1 | 0.571 | 63.333 | 93.75 | 10.133 | 0.391 | 0.849 (0.763, 0.913) |
| X2 | 0.379 | 81.667 | 56.25 | 1.867 | 0.326 | 0.728 (0.629, 0.812) |
| X3 | 0.271 | 85.833 | 41.25 | 1.461 | 0.343 | 0.684 (0.583, 0.773) |
Abbreviations: AUC, area under the curve; AUROC, area under the receiver operating characteristic; CI, confidence interval; LR, likelihood ratio.
Pairwise comparison of AUROC of diagnostic test X1, X2, and X3 for disease A
| Comparison | AUC difference | SE | ||
|---|---|---|---|---|
| X1 and X2 | 0.121 | 0.038 | 3.214 | 0.001 |
| X1 and X3 | 0.165 | 0.04 | 4.175 | < 0.001 |
| X2 and X3 | 0.044 | 0.048 | 0.912 | 0.362 |
Abbreviations: AUC, area under the curve; AUROC, area under the receiver operating characteristic; SE, standard error.
Observed and expected frequency of agreement among observer 1 and observer 2
| Observer 1 | Total | |||
|---|---|---|---|---|
| Diseased | Nondiseased | |||
|
|
| 86 (49) | 12 (51) | 98 |
|
| 14 (49) | 88 (51) | 102 | |
| Total | 100 | 100 | 200 | |
Data layout for a single diagnostic test using simple two-phase design
| Total | |||
|---|---|---|---|
|
|
|
|
|
|
|
|
|
|