OBJECTIVE: To evaluate the interrater reliability of rheumatologist diagnosis of rheumatoid arthritis (RA) and the concordance between rheumatologist and computer algorithms for assessing the accuracy of a diagnosis of RA. METHODS: Self-reported data regarding symptoms and signs for a diagnosis of RA were considered by a panel of rheumatologists and by computer algorithms to assess the probability of a diagnosis of RA for 90 patients. The rheumatologists' review was validated through medical record. RESULTS: The interrater reliability among rheumatologists regarding a diagnosis of RA was 84%; the chance-corrected agreement (kappa) was 0.66. Agreement between the rheumatologists' rating and the best-performing algorithm was 95%. Using rheumatologist's review as a standard, the sensitivity of the algorithm was 100%, specificity was 88%, and the positive predictive value was 91%. The validation of rheumatologist's review by medical record showed 81% sensitivity, 60% specificity, and 78% positive predictive value. CONCLUSION: Reliability of rheumatologists' assignment of a diagnosis of RA by using self-report data is good. Algorithms defining symptoms as either joint swelling or tenderness with symptom duration >or=4 weeks have a better agreement with rheumatologist's diagnosis than do ones relying on a longer symptom duration. RELEVANCE: These findings have important implications for health services research and quality improvement interventions pertinent to case finding for RA through self-report data.
OBJECTIVE: To evaluate the interrater reliability of rheumatologist diagnosis of rheumatoid arthritis (RA) and the concordance between rheumatologist and computer algorithms for assessing the accuracy of a diagnosis of RA. METHODS: Self-reported data regarding symptoms and signs for a diagnosis of RA were considered by a panel of rheumatologists and by computer algorithms to assess the probability of a diagnosis of RA for 90 patients. The rheumatologists' review was validated through medical record. RESULTS: The interrater reliability among rheumatologists regarding a diagnosis of RA was 84%; the chance-corrected agreement (kappa) was 0.66. Agreement between the rheumatologists' rating and the best-performing algorithm was 95%. Using rheumatologist's review as a standard, the sensitivity of the algorithm was 100%, specificity was 88%, and the positive predictive value was 91%. The validation of rheumatologist's review by medical record showed 81% sensitivity, 60% specificity, and 78% positive predictive value. CONCLUSION: Reliability of rheumatologists' assignment of a diagnosis of RA by using self-report data is good. Algorithms defining symptoms as either joint swelling or tenderness with symptom duration >or=4 weeks have a better agreement with rheumatologist's diagnosis than do ones relying on a longer symptom duration. RELEVANCE: These findings have important implications for health services research and quality improvement interventions pertinent to case finding for RA through self-report data.
Authors: David R Weir; Robert B Wallace; Kenneth M Langa; Brenda L Plassman; Robert S Wilson; David A Bennett; Ranjan Duara; David Loewenstein; Mary Ganguli; Mary Sano Journal: Alzheimers Dement Date: 2011-01 Impact factor: 21.566
Authors: Brian T Walitt; Florina Constantinescu; James D Katz; Arthur Weinstein; Hong Wang; Rohini K Hernandez; Judith Hsia; Barbara V Howard Journal: J Rheumatol Date: 2008-04-01 Impact factor: 4.666
Authors: Hannes Alder; Beat A Michel; Christian Marx; Giorgio Tamborrini; Thomas Langenegger; Pius Bruehlmann; Johann Steurer; Lukas M Wildi Journal: Int J Rheumatol Date: 2014-07-08