Literature DB >> 29677066

Agreement Analysis: What He Said, She Said Versus You Said.

Thomas R Vetter1, Patrick Schober2.   

Abstract

Correlation and agreement are 2 concepts that are widely applied in the medical literature and clinical practice to assess for the presence and strength of an association. However, because correlation and agreement are conceptually distinct, they require the use of different statistics. Agreement is a concept that is closely related to but fundamentally different from and often confused with correlation. The idea of agreement refers to the notion of reproducibility of clinical evaluations or biomedical measurements. The intraclass correlation coefficient is a commonly applied measure of agreement for continuous data. The intraclass correlation coefficient can be validly applied specifically to assess intrarater reliability and interrater reliability. As its name implies, the Lin concordance correlation coefficient is another measure of agreement or concordance. In undertaking a comparison of a new measurement technique with an established one, it is necessary to determine whether they agree sufficiently for the new to replace the old. Bland and Altman demonstrated that using a correlation coefficient is not appropriate for assessing the interchangeability of 2 such measurement methods. They in turn described an alternative approach, the since widely applied graphical Bland-Altman Plot, which is based on a simple estimation of the mean and standard deviation of differences between measurements by the 2 methods. In reading a medical journal article that includes the interpretation of diagnostic tests and application of diagnostic criteria, attention is conventionally focused on aspects like sensitivity, specificity, predictive values, and likelihood ratios. However, if the clinicians who interpret the test cannot agree on its interpretation and resulting typically dichotomous or binary diagnosis, the test results will be of little practical use. Such agreement between observers (interobserver agreement) about a dichotomous or binary variable is often reported as the kappa statistic. Assessing the interrater agreement between observers, in the case of ordinal variables and data, also has important biomedical applicability. Typically, this situation calls for use of the Cohen weighted kappa. Questionnaires, psychometric scales, and diagnostic tests are widespread and increasingly used by not only researchers but also clinicians in their daily practice. It is essential that these questionnaires, scales, and diagnostic tests have a high degree of agreement between observers. It is therefore vital that biomedical researchers and clinicians apply the appropriate statistical measures of agreement to assess the reproducibility and quality of these measurement instruments and decision-making processes.

Mesh:

Year:  2018        PMID: 29677066     DOI: 10.1213/ANE.0000000000002924

Source DB:  PubMed          Journal:  Anesth Analg        ISSN: 0003-2999            Impact factor:   5.108


  6 in total

1.  Reliability and Validity of the Malaysian English Version of the Diagnostic Criteria for Temporomandibular Disorder (M-English DC/TMD).

Authors:  Farah Nur Tedin Ng; Kathreena Kadir; Zamros Yuzadi Mohd Yusof
Journal:  Healthcare (Basel)       Date:  2022-02-09

2.  Comparison of 3D T1-SPACE and DSA in evaluation of intracranial in-stent restenosis.

Authors:  Qiuji Shao; Qiang Li; Qiaowei Wu; Tianxiao Li; Li Li; Kaitao Chang
Journal:  Br J Radiol       Date:  2020-12-01       Impact factor: 3.039

3.  Screening of ZIKA virus infection among dengue-like illness patients with negative RT-PCR for dengue virus in Punjab - Pakistan.

Authors:  Somia Iqtadar; Thuan Huu Vo; Mehreen Mehmood; Muhammad Salman; Mamunur Rahman Malik; Faisal Masud; Nhu Nguyen Tran Minh
Journal:  Pak J Med Sci       Date:  2021 May-Jun       Impact factor: 1.088

4.  The application of a workflow integrating the variable reproducibility and harmonizability of radiomic features on a phantom dataset.

Authors:  Abdalla Ibrahim; Turkey Refaee; Ralph T H Leijenaar; Sergey Primakov; Roland Hustinx; Felix M Mottaghy; Henry C Woodruff; Andrew D A Maidment; Philippe Lambin
Journal:  PLoS One       Date:  2021-05-07       Impact factor: 3.240

5.  MRI texture feature repeatability and image acquisition factor robustness, a phantom study and in silico study.

Authors:  Joshua Shur; Matthew Blackledge; James D'Arcy; David J Collins; Maria Bali; Martin O'Leach; Dow-Mu Koh
Journal:  Eur Radiol Exp       Date:  2021-01-19

6.  Repeated Measures Designs and Analysis of Longitudinal Data: If at First You Do Not Succeed-Try, Try Again.

Authors:  Patrick Schober; Thomas R Vetter
Journal:  Anesth Analg       Date:  2018-08       Impact factor: 5.108

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.