| Literature DB >> 34228319 |
Annette M Schmid1, David L Raunig2, Colin G Miller3, Richard C Walovitch4, Robert W Ford5, Michael O'Connor6, Guenther Brueggenwerth7, Josy Breuer8, Liz Kuney9, Robert R Ford10.
Abstract
The debate over human visual perception and how medical images should be interpreted have persisted since X-rays were the only imaging technique available. Concerns over rates of disagreement between expert image readers are associated with much of the clinical research and at times driven by the belief that any image endpoint variability is problematic. The deeper understanding of the reasons, value, and risk of disagreement are somewhat siloed, leading, at times, to costly and risky approaches, especially in clinical trials. Although artificial intelligence promises some relief from mistakes, its routine application for assessing tumors within cancer trials is still an aspiration. Our consortium of international experts in medical imaging for drug development research, the Pharma Imaging Network for Therapeutics and Diagnostics (PINTAD), tapped the collective knowledge of its members to ground expectations, summarize common reasons for reader discordance, identify what factors can be controlled and which actions are likely to be effective in reducing discordance. Reinforced by an exhaustive literature review, our work defines the forces that shape reader variability. This review article aims to produce a singular authoritative resource outlining reader performance's practical realities within cancer trials, whether they occur within a clinical or an independent central review.Entities:
Keywords: Clinical trials; Image interpretation; Independent review; Radiology; Reader disagreement; Visual perception
Mesh:
Year: 2021 PMID: 34228319 PMCID: PMC8259547 DOI: 10.1007/s43441-021-00316-6
Source DB: PubMed Journal: Ther Innov Regul Sci ISSN: 2168-4790 Impact factor: 1.778
A Sample of Evaluator Agreements for Different Specialties Since 1947.
| Author | Year | Disagreement or Error Rate | Type of Assessment |
|---|---|---|---|
| Birkelo et al. | 1947 | Inter-reader: 35% Intra-reader: 20% | 5 radiologists Tuberculosis radiological diagnosis Film |
| Thiesse et al. | 1997 | Major disagreements—40% Reasons: tumor measurements, selection of measurable targets, intercurrent diseases, and radiologic technical problems | Renal cell carcinoma Disagreement with committee of tumor response WHO criteria basis |
| Gwyther et al. | 1997 | Disagreement with response: 39% | Epithelial ovarian cancer Response: WHO criteria 2 Independent readers |
| Rubenfeld et al. | 1999 | 32% had ≥ 5 dissenters | Acute respiratory distress syndrome diagnosis (CT) 21 experts |
| Wormanns et al. | 2000 | 5 mm slice thickness Detection disagreement: 38% Size category: | Pulmonary nodules 2 readers Detection and size 23 patients 286 nodules |
| Aldape et al. | 2000 | Disagreement: 23% | Glioma Digital pathology/neuropathology Diagnosis |
| Pandolfino et al. | 2002 | Intra-observer Experts: 0.55 Trainees: 0.44 Inter-observer Experts: 0.56 Trainees: 0.46 | Endoscopic scoring of esophagitis Experts and trainees |
| Scholten et al. | 2004 | FIGO disagreement: 30% ( | Endometrial carcinoma Digital Histology (FIGO) 2 independent pathologists |
| Gietema et al. | 2006 | Discrepant volumes: 10.9% Inter-reader Spearman Correlation: r = 0.99 | Lung cancer nodule detection ( Local and Central reader Volume |
| Hricak H, et al. | 2007 | Staging CT: Visualization CT: Sens/Spec Sens: CT = .26 MRI = .48 Spec: CT = .92 MRI = .79 | Cervical cancer Diagnosis 4 radiologists (CT) 4 radiologists (MRI) |
| Hersh et al. | 2007 | All combinations of readers Disagreement = 25% | Lobe- predominant emphysema HRCT Pulmonologists and radiologists |
| Suzuki et al. | 2010 | Inter-reader agreement 0.72) Intra-reader agreement 0.96) | Breast and Colorectal cancer RECIST response 2 radiologists |
| Ibrahim et al. | 2011 | Inter-reader agreement for subarachnoid hemorrhage | Aneurysmal subarachnoid hemorrhage 1 neurosurgeon 1 neuroradiologist |