| Literature DB >> 26229664 |
Josephine C Debono1, Ann E Poulos2.
Abstract
INTRODUCTION: The aim of this study was to first evaluate the quality of studies investigating the diagnostic accuracy of radiographers as mammogram screen-readers and then to develop an adapted tool for determining the quality of screen-reading studies.Entities:
Keywords: Accuracy; evaluation tools; mammogram; quality; radiographers; screen-readers
Year: 2014 PMID: 26229664 PMCID: PMC4364803 DOI: 10.1002/jmrs.68
Source DB: PubMed Journal: J Med Radiat Sci ISSN: 2051-3895
Classification of items included in quality assessment tools (Source, with permission: Whiting et al.17 p. 3, © 2005, Elsevier) plus observer characteristics (Source, with permission: Brealey and Westwood21 p. 676, © 2006, the British Institute of Radiology)
| ID | Item | Description of item |
|---|---|---|
| A. Potential for bias | ||
| A1 | Reference standard | Was an appropriate reference standard used to determine the presence or absence of the target condition? |
| A2 | Disease progression bias | Could a change in disease state have occurred between application of the index test and reference standard? |
| A3 | Verification bias | Did all subjects receive verification of the target condition using the same reference standard? |
| A4 | Incorporation bias | Did the index test form part of the reference test? |
| A5 | Treatment paradox | Was treatment started based on the result of the index test before the reference standard was applied? |
| A6 | Review bias | Were index test results interpreted without knowledge of the results of the reference standard, and vice versa? |
| A7 | Clinical review bias | Was clinical information available when test results were interpreted? |
| A8 | Observer/instrument variation | Was observer/instrument variation likely to have affected estimates of test performance? |
| A9 | Handling of uninterpretable results | Were uninterpretable results included in the analysis? |
| A10 | Arbitrary choice of threshold value | Was the threshold value chosen independently of the results of the study? i.e., it should not have been chosen to optimise estimates of test performance |
| B. Applicability | ||
| B1 | Spectrum composition | Was the population studied similar to the one in which you are interested? |
| B2 | Population recruitment | Was the method of population recruitment adequate to include an appropriate spectrum of patients? |
| B3 | Disease prevalence/severity | Was the spectrum of disease prevalence and severity similar to the one in which you are interested? |
| B4 | Change in technology of index test | Is it likely that the technology of the test has changed since the study was conducted? |
| C. Conduct of the study | ||
| C1 | Subgroup analysis | Were subgroup analyses appropriate and specified? |
| C2 | Sample size | Were an appropriate number of participants included in the study? |
| C3 | Objectives | Were study objectives relevant to the study question? |
| C4 | Protocol | Was a study protocol developed before the study started and did the investigators adhere to it? |
| D. Reporting of the study | ||
| D1 | Inclusion criteria | Were inclusion criteria clearly reported? |
| D2 | Test execution | Were sufficient details provided on how the index test was performed to permit its replication? |
| D3 | Reference execution | Were sufficient details provided on how the reference standard was performed to permit its replication? |
| D4 | Normal defined | Did the authors clearly report what they considered to be a normal test result? |
| D5 | Appropriate results | Were appropriate results presented? e.g., sensitivity, specificity, likelihood ratios |
| D6 | Precision of results | Was some estimate of the precision of the results presented? e.g., confidence interval |
| D7 | Drop-outs | Were all patients that entered the study accounted for? |
| D8 | Data table | Was an |
| D9 | Utility of test | Was there some indication of how useful the test might be in practice? |
| E. Observer characteristics | ||
| E1 | Image allocation to observers | How were images allocated to be read by the observers? |
| E2 | Number of observers | Was the number of observers presented? |
| E3 | Observer experience | Was the experience of the observers described? |
| E4 | Observer training | Was the training of the observers described? |
| E5 | Observer profession | Was the profession of the observers presented? |
| E6 | Observer variability | Was there an assessment of observer variability? |
| E7 | Analysis of observer variability | Was observer variability considered in the analyses of test accuracy? |
Screen-reading studies, in chronological order
| Authors | Title |
|---|---|
| Haiart and Henderson | A comparison of interpretation of screening mammograms by a radiographer, a doctor and a radiologist |
| Bassett et al. | Effects of a program to train radiologic technologists to identify abnormalities on mammograms |
| Pauli et al. | Comparison of radiographer/radiologist double film reading with single reading in breast cancer screening |
| Pauli et al. | Radiographers as film readers in screening mammography: an assessment of competence under test and screening conditions |
| Tonita et al. | Medical radiologic technologist review: effects on a population-based breast cancer screening program |
| Wivell et al. | Can radiographers read screening mammograms? |
| Sumkin et al. | Prescreening mammography by technologists: a preliminary assessment |
| Holt | Evaluating radiological technologists' ability to detect abnormalities in film-screen mammographic images: A decision analysis pilot project |
| Duijm et al. | Additional double reading of screening mammograms by radiologic technologists: impact on screening performance parameters |
| Duijm et al. | Introduction of additional double reading of mammograms by radiographers: effects on a biennial screening programme outcome |
| Duijm et al. | Inter-observer variability in mammography screening and effect of type and number of readers on screening outcome |
Evaluation of reviewed studies using the constructed quality tool (Table1)
| Study | Haiart and Henderson | Bassett et al. | Pauli et al. | Pauli et al. | Tonita et al. | Wivell et al. | Sumkin et al. | Holt | Duijm et al. | Duijm et al. | Duijm et al. | Total |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| A. Potential for bias | ||||||||||||
| A1 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | – | ✓ | ✓ | ✓ | ✓ | 1 |
| A2 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | – | ✓ | ✓ | ✓ | ✓ | 1 |
| A3 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | 0 |
| A4 | – | ✓ | – | Partial | – | ✓ | – | ✓ | – | – | – | 7.5 |
| A5 | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A |
| A6 | ✓ | ✓ | ✓ | ✓ | Partial | ✓ | ✓ | ✓ | ✓ | ✓ | Partial | 0 |
| A7 | N/S | – | ✓ | ✓ | N/S | ✓ | ✓ | – | ✓ | ✓ | ✓ | 2 |
| A8 | – | – | – | – | – | – | – | – | – | – | – | 11 |
| A9 | – | – | – | – | – | – | – | – | – | – | – | 11 |
| A10 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | 0 |
| Total | 3 | 2 | 3 | 2.5 | 3.5 | 2 | 5 | 3 | 3 | 3 | 3.5 | 33.5 |
| B. Applicability of results | ||||||||||||
| B1 | ✓ | N/S | ✓ | ✓ | ✓ | ✓ | ✓ | – | ✓ | ✓ | ✓ | 1 |
| B2 | ✓ | N/S | ✓ | ✓ | ✓ | ✓ | – | – | ✓ | ✓ | ✓ | 2 |
| B3 | ✓ | – | ✓ | ✓ | ✓ | ✓ | – | – | ✓ | ✓ | ✓ | 3 |
| B4 | – | – | – | – | – | – | – | – | – | – | – | 11 |
| Total | 1 | 2 | 1 | 1 | 1 | 1 | 3 | 4 | 1 | 1 | 1 | 17 |
| C. Conduct of the study | ||||||||||||
| C1 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | 0 |
| C2 | – | ✓ | ✓ | ✓ | – | – | ✓ | – | ✓ | ✓ | ✓ | 4 |
| C3 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | 0 |
| C4 | N/S | N/S | N/S | N/S | N/S | N/S | N/S | N/S | N/S | N/S | N/S | 0 |
| Total | 1 | 0 | 0 | 0 | 1 | 1 | 0 | 1 | 0 | 0 | 0 | 4 |
| D. Reporting of the study | ||||||||||||
| D1 | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A | N/A |
| D2 | Partial | ✓ | ✓ | ✓ | Partial | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | 1 |
| D3 | ✓ | ✓ | ✓ | – | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | 1 |
| D4 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | 0 |
| D5 | ✓ | ✓ | ✓ | ✓ | – | – | – | ✓ | ✓ | – | ✓ | 4 |
| D6 | – | – | – | – | ✓ | – | – | – | ✓ | – | ✓ | 8 |
| D7 | – | – | ✓ | – | – | – | – | – | – | – | – | 10 |
| D8 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | 0 |
| D9 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | 0 |
| Total | 2.5 | 2 | 1 | 3 | 2.5 | 3 | 3 | 2 | 1 | 3 | 1 | 24 |
| E. Observer characteristics | ||||||||||||
| E1 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | – | ✓ | ✓ | ✓ | ✓ | 1 |
| E2 | ✓ | ✓ | Partial | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | 5 |
| E3 | – | – | ✓ | ✓ | – | – | ✓ | ✓ | ✓ | ✓ | ✓ | 4 |
| E4 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | 2 |
| E5 | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | ✓ | 0 |
| E6 | ✓ | ✓ | – | – | – | – | – | ✓ | – | – | ✓ | 7 |
| E7 | – | – | – | – | – | ✓ | – | – | – | – | – | 10 |
| Total | 2 | 2 | 2.5 | 2 | 3 | 2 | 4 | 2 | 2 | 2 | 1 | 24.5 |
N/S, not stated; N/A, not applicable.
Developed tool named DASQUART (Diagnostic Accuracy Study QUality And Reporting Tool) for determining quality in studies investigating diagnostic accuracy in screen-reading
| Criteria | Description of criteria | |
|---|---|---|
| A1 | Reference standard | An appropriate reference standard of pathology and at least 1 year follow-up used to determine the presence or absence of breast cancer |
| A2 | Disease Progression bias | An interval cancer could not occur between the initial mammogram and the reference standard |
| A3 | Verification bias | Same reference standard applied across the study |
| A4 | Incorporation bias | The reading of the screening mammogram does not form part of the reference standard |
| A6 | Review bias | Mammograms read blinded to knowledge of reference standard and interpretation by other readers |
| A7 | Clinical review bias | Previous image rounds available for comparison |
| A8 | Instrument variation | No reporting instrument variation which will affect estimates of test performance, e.g., use of BIRADS® lexicon |
| A9 | Handling of uninterpretable results | Uninterpretable results included in the analysis |
| A10 | Arbitrary choice of threshold value | Threshold value of normal chosen independently of results |
| B1 | Spectrum composition | Image sample similar to one of interest (test sets, e.g., PERFORMS, |
| B2 | Population recruitment | Image sample selected adequate to include appropriate spectrum (test sets, e.g., PERFORMS, |
| B3 | Disease prevalence/severity | Spectrum of breast cancer prevalence similar to one of interest (test sets, e.g., PERFORMS, |
| B4 | Change in technology of index test | No change in mammography technology which will affect applicability of results |
| C1 | Subgroup analysis | Subgroup analyses were appropriate and specified |
| C2 | Sample size | Appropriate number of images included in study |
| C3 | Objectives | Study objectives relevant to study question |
| C4 | Study design | The purpose, method, results and conclusions demonstrate logical coherence and consistency |
| D1 | Inclusion criteria | Included in systematic reviews |
| D2 | Test execution (a) images | Sufficient details of mammogram reading reported to permit its replication. Details include number of images read in total and at one sitting, how images were selected (test sets), degree of difficulty (test sets), types of breast cancers included (test sets). |
| D2 | Test execution (b) environment | Time taken to read, background lighting and type of monitors |
| D3 | Reference execution | Sufficient details provided of reference standard used to permit its replication |
| D4 | Normal defined | Authors clearly reported what was considered a normal reading result |
| D5 | Appropriate results | Appropriate results of accuracy presented, e.g., sensitivity, specificity, ROC and JAFROC analysis |
| D6 | Precision of results | Estimate of precision of results presented as appropriate |
| D7 | Drop-outs | All images and observers accounted for |
| D8 | Data table | Test performance reported in a data table |
| D9 | Utility of test | Clinical relevance of the test emphasised |
| E1 | Image allocation to observers | Image allocation to observers described |
| E2 | Number of observers | Number of observers presented |
| E3 | Observer experience | Experience of observers described |
| E4 | Observer training | Training of observers described |
| E5 | Observer profession | Profession of observers presented |
| E6 | Analysis of observer variability | Observer variability in analysis, e.g., Kappa statistic |