Literature DB >> 34305250

Methods of assessing categorical agreement between correlated screening tests in clinical studies.

Thomas J Zhou1, Sughra Raza2, Kerrie P Nelson1.   

Abstract

Advances in breast imaging and other screening tests have prompted studies to evaluate and compare the consistency between experts' ratings of existing with new screening tests. In clinical settings, medical experts make subjective assessments of screening test results such as mammograms. Consistency between experts' ratings is evaluated by measures of inter-rater agreement or association. However, conventional measures, such as Cohen's and Fleiss' kappas, are unable to be applied or may perform poorly when studies consist of many experts, unbalanced data, or dependencies between experts' ratings exist. Here we assess the performance of existing approaches including recently developed summary measures for assessing the agreement between experts' binary and ordinal ratings when patients undergo two screening procedures. Methods to assess consistency between repeated measurements by the same experts are also described. We present applications to three large-scale clinical screening studies. Properties of these agreement measures are illustrated via simulation studies. Generally, a model-based approach provides several advantages over alternative methods including the ability to flexibly incorporate various measurement scales (i.e. binary or ordinal), large numbers of experts and patients, sparse data, and robustness to prevalence of underlying disease.

Entities:  

Keywords:  Fleiss’ kappa; Intra-rater agreement; association; binary ratings; ordinal classifications; rater training

Year:  2020        PMID: 34305250      PMCID: PMC8299998          DOI: 10.1080/02664763.2020.1777394

Source DB:  PubMed          Journal:  J Appl Stat        ISSN: 0266-4763            Impact factor:   1.404


  23 in total

1.  2 x 2 kappa coefficients: measures of agreement or association.

Authors:  D A Bloch; H C Kraemer
Journal:  Biometrics       Date:  1989-03       Impact factor: 2.571

2.  American College of Radiology Imaging Network digital mammographic imaging screening trial: objectives and methodology.

Authors:  Etta D Pisano; Constantine A Gatsonis; Martin J Yaffe; R Edward Hendrick; Anna N A Tosteson; Dennis G Fryback; Lawrence W Bassett; Janet K Baum; Emily F Conant; Roberta A Jong; Murray Rebner; Carl J D'Orsi
Journal:  Radiology       Date:  2005-06-16       Impact factor: 11.105

3.  Accuracy of soft-copy digital mammography versus that of screen-film mammography according to digital manufacturer: ACRIN DMIST retrospective multireader study.

Authors:  R Edward Hendrick; Elodia B Cole; Etta D Pisano; Suddhasatta Acharyya; Helga Marques; Michael A Cohen; Roberta A Jong; Gordon E Mawdsley; Kalpana M Kanal; Carl J D'Orsi; Murray Rebner; Constantine Gatsonis
Journal:  Radiology       Date:  2008-04       Impact factor: 11.105

4.  Weighted kappa: nominal scale agreement with provision for scaled disagreement or partial credit.

Authors:  J Cohen
Journal:  Psychol Bull       Date:  1968-10       Impact factor: 17.737

5.  Comparison of full-field digital mammography with screen-film mammography for cancer detection: results of 4,945 paired examinations.

Authors:  J M Lewin; R E Hendrick; C J D'Orsi; P K Isaacs; L J Moss; A Karellas; G A Sisney; C C Kuni; G R Cutter
Journal:  Radiology       Date:  2001-03       Impact factor: 11.105

Review 6.  Misinterpretation and misuse of the kappa statistic.

Authors:  M Maclure; W C Willett
Journal:  Am J Epidemiol       Date:  1987-08       Impact factor: 4.897

7.  The measurement of observer agreement for categorical data.

Authors:  J R Landis; G G Koch
Journal:  Biometrics       Date:  1977-03       Impact factor: 2.571

8.  Diagnostic performance of digital versus film mammography for breast-cancer screening.

Authors:  Etta D Pisano; Constantine Gatsonis; Edward Hendrick; Martin Yaffe; Janet K Baum; Suddhasatta Acharyya; Emily F Conant; Laurie L Fajardo; Lawrence Bassett; Carl D'Orsi; Roberta Jong; Murray Rebner
Journal:  N Engl J Med       Date:  2005-09-16       Impact factor: 91.245

9.  Training Programs on Endoscopic Scoring Systems for Inflammatory Bowel Disease Lead to a Significant Increase in Interobserver Agreement Among Community Gastroenterologists.

Authors:  Marco Daperno; Michele Comberlato; Fabrizio Bossa; Alessandro Armuzzi; Livia Biancone; Andrea G Bonanomi; Andrea Cassinotti; Rocco Cosintino; Giovanni Lombardi; Roberto Mangiarotti; Alfredo Papa; Roberta Pica; Luca Grassano; Guido Pagana; Renata D'Incà; Ambrogio Orlando; Fernando Rizzello
Journal:  J Crohns Colitis       Date:  2017-05-01       Impact factor: 9.071

10.  A measure of association for ordered categorical data in population-based studies.

Authors:  Kerrie P Nelson; Don Edwards
Journal:  Stat Methods Med Res       Date:  2016-05-16       Impact factor: 3.021

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.