Literature DB >> 31099902

A paired kappa to compare binary ratings across two medical tests.

Kerrie P Nelson1, Don Edwards2.   

Abstract

Agreement between experts' ratings is an important prerequisite for an effective screening procedure. In clinical settings, large-scale studies are often conducted to compare the agreement of experts' ratings between new and existing medical tests, for example, digital versus film mammography. Challenges arise in these studies where many experts rate the same sample of patients undergoing two medical tests, leading to a complex correlation structure between experts' ratings. Here, we propose a novel paired kappa measure to compare the agreement between the binary ratings of many experts across two medical tests. Existing approaches can accommodate only a small number of experts, rely heavily on Cohen's kappa and Scott's pi measures of agreement, and thus are prone to their drawbacks. The proposed kappa appropriately accounts for correlations between ratings due to patient characteristics, corrects for agreement due to chance, and is robust to disease prevalence and other flaws inherent in the use of Cohen's kappa. It can be easily calculated in the software package R. In contrast to existing approaches, the proposed measure can flexibly incorporate large numbers of experts and patients by utilizing the generalized linear mixed models framework. It is intended to be used in population-based studies, increasing efficiency without increasing modeling complexity. Extensive simulation studies demonstrate low bias and excellent coverage probability of the proposed kappa under a broad range of conditions. Methods are applied to a recent nationwide breast cancer screening study comparing film mammography to digital mammography.
© 2019 John Wiley & Sons, Ltd.

Entities:  

Keywords:  agreement; binary classifications; breast imaging; kappa; screening test

Year:  2019        PMID: 31099902      PMCID: PMC6884009          DOI: 10.1002/sim.8200

Source DB:  PubMed          Journal:  Stat Med        ISSN: 0277-6715            Impact factor:   2.373


  29 in total

1.  2 x 2 kappa coefficients: measures of agreement or association.

Authors:  D A Bloch; H C Kraemer
Journal:  Biometrics       Date:  1989-03       Impact factor: 2.571

2.  Latent-variable models for longitudinal data with bivariate ordinal outcomes.

Authors:  David Todem; KyungMann Kim; Emmanuel Lesaffre
Journal:  Stat Med       Date:  2007-02-28       Impact factor: 2.373

3.  Mammography: interobserver variability in breast density assessment.

Authors:  E A Ooms; H M Zonderland; M J C Eijkemans; M Kriege; B Mahdavian Delavary; C W Burger; A C Ansink
Journal:  Breast       Date:  2007-12       Impact factor: 4.380

4.  Adjusted inference procedures for the interobserver agreement in twin studies.

Authors:  Stephanie N Dixon; Allan Donner; Mohamed M Shoukri
Journal:  Stat Methods Med Res       Date:  2013-03-12       Impact factor: 3.021

Review 5.  Estimating diagnostic accuracy without a gold standard: A continued controversy.

Authors:  John Collins; Paul S Albert
Journal:  J Biopharm Stat       Date:  2016-08-22       Impact factor: 1.051

6.  A random-effects ordinal regression model for multilevel analysis.

Authors:  D Hedeker; R D Gibbons
Journal:  Biometrics       Date:  1994-12       Impact factor: 2.571

7.  Estimating the error rates of diagnostic tests.

Authors:  S L Hui; S D Walter
Journal:  Biometrics       Date:  1980-03       Impact factor: 2.571

8.  Diagnostic performance of digital versus film mammography for breast-cancer screening.

Authors:  Etta D Pisano; Constantine Gatsonis; Edward Hendrick; Martin Yaffe; Janet K Baum; Suddhasatta Acharyya; Emily F Conant; Laurie L Fajardo; Lawrence Bassett; Carl D'Orsi; Roberta Jong; Murray Rebner
Journal:  N Engl J Med       Date:  2005-09-16       Impact factor: 91.245

9.  Breast cancer screening using tomosynthesis in combination with digital mammography.

Authors:  Sarah M Friedewald; Elizabeth A Rafferty; Stephen L Rose; Melissa A Durand; Donna M Plecha; Julianne S Greenberg; Mary K Hayes; Debra S Copit; Kara L Carlson; Thomas M Cink; Lora D Barke; Linda N Greer; Dave P Miller; Emily F Conant
Journal:  JAMA       Date:  2014-06-25       Impact factor: 56.272

10.  An interobserver reproducibility study on invasiveness of bladder cancer using virtual microscopy and heatmaps.

Authors:  Eva Compérat; Lars Egevad; Antonio Lopez-Beltran; Philippe Camparo; Ferran Algaba; Mahul Amin; Jonathan I Epstein; Hans Hamberg; Christina Hulsbergen-van de Kaa; Glen Kristiansen; Rodolfo Montironi; Chin-Chen Pan; Fabrice Heloir; Kilian Treurniet; Jenna Sykes; Theo H Van der Kwast
Journal:  Histopathology       Date:  2013-09-16       Impact factor: 5.087

View more
  1 in total

1.  Methods of assessing categorical agreement between correlated screening tests in clinical studies.

Authors:  Thomas J Zhou; Sughra Raza; Kerrie P Nelson
Journal:  J Appl Stat       Date:  2020-06-09       Impact factor: 1.404

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.