Literature DB >> 26229664

Evaluating radiographers' diagnostic accuracy in screen-reading mammograms: what constitutes a quality study?

Josephine C Debono1, Ann E Poulos2.   

Abstract

INTRODUCTION: The aim of this study was to first evaluate the quality of studies investigating the diagnostic accuracy of radiographers as mammogram screen-readers and then to develop an adapted tool for determining the quality of screen-reading studies.
METHODS: A literature search was used to identify relevant studies and a quality evaluation tool constructed by combining the criteria for quality of Whiting, Rutjes, Dinnes et al. and Brealey and Westwood. This constructed tool was then applied to the studies and subsequently adapted specifically for use in evaluating quality in studies investigating diagnostic accuracy of screen-readers.
RESULTS: Eleven studies were identified and the constructed tool applied to evaluate quality. This evaluation resulted in the identification of quality issues with the studies such as potential for bias, applicability of results, study conduct, reporting of the study and observer characteristics. An assessment of the applicability and relevance of the tool for this area of research resulted in adaptations to the criteria and the development of a tool specifically for evaluating diagnostic accuracy in screen-reading.
CONCLUSIONS: This tool, with further refinement and rigorous validation can make a significant contribution to promoting well-designed studies in this important area of research and practice.

Entities:  

Keywords:  Accuracy; evaluation tools; mammogram; quality; radiographers; screen-readers

Year:  2014        PMID: 26229664      PMCID: PMC4364803          DOI: 10.1002/jmrs.68

Source DB:  PubMed          Journal:  J Med Radiat Sci        ISSN: 2051-3895


Introduction

Diagnostic accuracy in medical imaging is essential for appropriate patient management and treatment.1 Accurate screen-reading of mammogram images is critical for the early detection of breast cancer, the goal of population screening programs.1 Screen-readers of mammogram images are predominantly, but not exclusively, radiologists.2 Currently, there are workforce issues in radiology which impact on their availability for screen-reading.3 In the United Kingdom, this shortage has been addressed by the training and employment of radiographers as screen-readers.2,4,5 A range of studies have investigated the diagnostic accuracy of radiographers in this role,2,6–15 and provide evidence that radiographers have comparable accuracy to radiologists.2,6–15 More recent studies provide evidence of the ability of radiographers to contribute to improvements in the efficiency of the screening process, and most importantly that combining radiologist and radiographer screen-reading has been found to improve cancer detection rates.7–9 Since diagnostic accuracy in screen-reading underpins the goal of breast screening to detect breast cancer early and reduce mortality, the quality of these studies is paramount. A systematic review published in 2008 by van den Biggelaar et al.16 excluded articles without evidence of sensitivity and specificity and an appropriate gold standard, resulting in a total of six. This systematic review raised questions of what constitutes a well-designed study and how quality is defined in studies investigating screen-reading accuracy by radiographers. More specifically, the authors emphasised the necessity of determining the key components of a well-designed study in this area of research to increase the rigour and applicability of the outcomes to the clinical environment.

Quality evaluation tools for studies of diagnostic accuracy

A number of tools for evaluating the quality of diagnostic accuracy studies have been identified in the literature.17 The Standards for the Reporting of Diagnostic accuracy studies (STARD), was developed from an initiative to improve the accuracy and completeness of reporting studies of diagnostic accuracy.18 The Quality of Diagnostic Accuracy Studies (QUADAS) tool was later developed and validated by Whiting et al.19 to determine the quality of primary studies in systematic reviews of diagnostic accuracy. Subsequently, Whiting, et al.17 conducted a systematic review of existing quality assessment tools to examine both the extent and type of quality assessment being incorporated in diagnostic accuracy systematic reviews. Aspects of quality considered in their review were classified as: potential for bias; conduct of the study; applicability of the results; and quality of reporting. Following data extraction, the data were synthesised according to purpose and summarised as items.19 This classification is useful since it is all-inclusive and includes items of quality drawn from an extensive review of systematic reviews. As well as determining the individual items related to quality in diagnostic accuracy studies, the classification also synthesises these items into aspects of quality. Importantly, this classification includes quality items relating to the reporting of studies.20 The comprehensive nature of this classification facilitates the adaption of the quality items or criteria to a specific area of diagnostic accuracy research such as screen-reading.

Importance of observer characteristics and variability

The importance of observer characteristics and variability on diagnostic accuracy in medical imaging have been emphasised by Brealey and Westwood,21 who claim that observers are frequently ignored in diagnostic accuracy studies in medical imaging in spite of their ability to affect the study outcomes. The number of observers, for example, influences the internal and external validity of research studies, while the profession and experience of observers affect estimates of accuracy.22,23 Brealey and Westwood21 strongly recommend the inclusion of observer assessment criteria in a quality assessment tool evaluating diagnostic accuracy in medical imaging. The aim of this study was firstly to evaluate the quality of the studies investigating the diagnostic accuracy of radiographers as screen-readers using a quality evaluation tool constructed by combining the criteria for quality of Whiting et al.17 and Brealey and Westwood.21 Secondly, the applicability and appropriateness of the criteria were determined and an adapted quality evaluation tool was developed specifically for use in evaluating diagnostic the accuracy in screen-reading studies.

Method

Stage 1: Quality evaluation tool for studies using imaging

To construct this quality evaluation tool the classifications and items of Whiting et al.17 were combined with the observer characteristics recommended by Brealey and Westwood21 to provide a comprehensive all-inclusive quality assessment tool for diagnostic accuracy studies using imaging. An Ethics statement is not applicable to this study.

Stage 2: Literature search

A literature search was undertaken within the Medline, PubMed, Web of Science and Cinahl databases, using combinations of the terms: mammogram, radiographer, technologist, screen-reading, accuracy and interpretation. This search was undertaken in 2010, and therefore limited to articles published at that time. An initial review of titles and abstracts enabled the exclusion of papers that were clearly not relevant to the subject of interest. Studies investigating the diagnostic accuracy of radiographers reading mammograms were selected. Further studies were located using the reference lists. As only a small number (n = 11) of papers were located, no further inclusion/exclusion criteria were applied.

Stage 3: Quality evaluation of reviewed studies

The quality evaluation of studies was carried out by two experienced researchers. The role of these researchers was firstly to adapt the ‘generic’ diagnostic accuracy study quality items in Table1 to specific criteria of quality in radiographer mammography screen-reading studies. This required knowledge and experience in mammography and the diagnostic process of screen-reading. Secondly, these researchers required research skills and experience in critical analysis of the reviewed studies to determine the extent to which they complied with the quality criteria. Finally, knowledge and familiarity with current relevant literature was required for stage 4. Any variation between the researchers was dealt with by discussion and consensus.
Table 1

Classification of items included in quality assessment tools (Source, with permission: Whiting et al.17 p. 3, © 2005, Elsevier) plus observer characteristics (Source, with permission: Brealey and Westwood21 p. 676, © 2006, the British Institute of Radiology)

IDItemDescription of item
A. Potential for bias
 A1Reference standardWas an appropriate reference standard used to determine the presence or absence of the target condition?
 A2Disease progression biasCould a change in disease state have occurred between application of the index test and reference standard?
 A3Verification biasDid all subjects receive verification of the target condition using the same reference standard?
 A4Incorporation biasDid the index test form part of the reference test?
 A5Treatment paradoxWas treatment started based on the result of the index test before the reference standard was applied?
 A6Review biasWere index test results interpreted without knowledge of the results of the reference standard, and vice versa?
 A7Clinical review biasWas clinical information available when test results were interpreted?
 A8Observer/instrument variationWas observer/instrument variation likely to have affected estimates of test performance?
 A9Handling of uninterpretable resultsWere uninterpretable results included in the analysis?
 A10Arbitrary choice of threshold valueWas the threshold value chosen independently of the results of the study? i.e., it should not have been chosen to optimise estimates of test performance
B. Applicability
 B1Spectrum compositionWas the population studied similar to the one in which you are interested?
 B2Population recruitmentWas the method of population recruitment adequate to include an appropriate spectrum of patients?
 B3Disease prevalence/severityWas the spectrum of disease prevalence and severity similar to the one in which you are interested?
 B4Change in technology of index testIs it likely that the technology of the test has changed since the study was conducted?
C. Conduct of the study
 C1Subgroup analysisWere subgroup analyses appropriate and specified?
 C2Sample sizeWere an appropriate number of participants included in the study?
 C3ObjectivesWere study objectives relevant to the study question?
 C4ProtocolWas a study protocol developed before the study started and did the investigators adhere to it?
D. Reporting of the study
 D1Inclusion criteriaWere inclusion criteria clearly reported?
 D2Test executionWere sufficient details provided on how the index test was performed to permit its replication?
 D3Reference executionWere sufficient details provided on how the reference standard was performed to permit its replication?
 D4Normal definedDid the authors clearly report what they considered to be a normal test result?
 D5Appropriate resultsWere appropriate results presented? e.g., sensitivity, specificity, likelihood ratios
 D6Precision of resultsWas some estimate of the precision of the results presented? e.g., confidence interval
 D7Drop-outsWere all patients that entered the study accounted for?
 D8Data tableWas an n x n table of test performance reported?
 D9Utility of testWas there some indication of how useful the test might be in practice?
E. Observer characteristics
 E1Image allocation to observersHow were images allocated to be read by the observers?
 E2Number of observersWas the number of observers presented?
 E3Observer experienceWas the experience of the observers described?
 E4Observer trainingWas the training of the observers described?
 E5Observer professionWas the profession of the observers presented?
 E6Observer variabilityWas there an assessment of observer variability?
 E7Analysis of observer variabilityWas observer variability considered in the analyses of test accuracy?
Classification of items included in quality assessment tools (Source, with permission: Whiting et al.17 p. 3, © 2005, Elsevier) plus observer characteristics (Source, with permission: Brealey and Westwood21 p. 676, © 2006, the British Institute of Radiology)

Stage 4: Development of quality assessment tool for mammography screen-reading

Following the process of evaluation, the criteria were adapted to the specific quality aspects of studies reporting on screen-reading. Adaptations to the criteria were identified that increased the relevance and applicability of the tool for the specific purpose of the evaluation of the diagnostic accuracy of screen-readers interpreting mammograms in breast screening facilities. A search of relevant literature was carried out for evidence of specific quality criteria.

Results and Discussion

Stage 1: Development of quality evaluation tool

The developed tool is presented in Table1. Eleven studies were identified in the literature relating to the diagnostic accuracy of radiographers reading screening mammograms and are presented in Table2. No studies were excluded from the review.
Table 2

Screen-reading studies, in chronological order

AuthorsTitle
Haiart and Henderson10A comparison of interpretation of screening mammograms by a radiographer, a doctor and a radiologist
Bassett et al.6Effects of a program to train radiologic technologists to identify abnormalities on mammograms
Pauli et al.2Comparison of radiographer/radiologist double film reading with single reading in breast cancer screening
Pauli et al.12Radiographers as film readers in screening mammography: an assessment of competence under test and screening conditions
Tonita et al.14Medical radiologic technologist review: effects on a population-based breast cancer screening program
Wivell et al.15Can radiographers read screening mammograms?
Sumkin et al.13Prescreening mammography by technologists: a preliminary assessment
Holt11Evaluating radiological technologists' ability to detect abnormalities in film-screen mammographic images: A decision analysis pilot project
Duijm et al.7Additional double reading of screening mammograms by radiologic technologists: impact on screening performance parameters
Duijm et al.8Introduction of additional double reading of mammograms by radiographers: effects on a biennial screening programme outcome
Duijm et al.9Inter-observer variability in mammography screening and effect of type and number of readers on screening outcome
Screen-reading studies, in chronological order Study quality of each of the 11 studies was evaluated using the developed tool (stage 2 of method); the results of these evaluations are presented in Table3.
Table 3

Evaluation of reviewed studies using the constructed quality tool (Table1)

StudyHaiart and Henderson10Bassett et al.6Pauli et al.2Pauli et al.12Tonita et al.14Wivell et al.15Sumkin et al.13Holt11Duijm et al.7Duijm et al.8Duijm et al.9Total
A. Potential for bias
A11
A21
A30
A4Partial7.5
A5N/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/A
A6PartialPartial0
A7N/SN/S2
A811
A911
A100
Total3232.53.5253333.533.5
B. Applicability of results
B1N/S1
B2N/S2
B33
B411
Total1211113411117
C. Conduct of the study
C10
C24
C30
C4N/SN/SN/SN/SN/SN/SN/SN/SN/SN/SN/S0
Total100011010004
D. Reporting of the study
D1N/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/A
D2PartialPartial1
D31
D40
D54
D68
D710
D80
D90
Total2.52132.533213124
E. Observer characteristics
E11
E2Partial5
E34
E42
E50
E67
E710
Total222.52324222124.5

N/S, not stated; N/A, not applicable.

Evaluation of reviewed studies using the constructed quality tool (Table1) N/S, not stated; N/A, not applicable. The ‘Total’ row under each category A–E indicates the numbers of negative responses to the criteria for each study, while the numbers in the final ‘Total’ column indicate the number of negative responses to each of the 34 criteria. If a partial negative response was indicated then 0.5 was allocated. This relates positively to the scoring method used by Whiting et al.24 The ‘scoring’ of quality is fraught with difficulties in interpretation. Whiting et al.25 emphasised the need to investigate individual quality items and their association with estimates of diagnostic accuracy rather than produce scores. So while identification of negative responses to criteria is a simplistic method of scoring quality, for the purposes of this study this method has provided detail of the categories demonstrating a large number of negative responses. The largest number of negative responses is identified in section A: Potential for bias. Potential bias can severely compromise outcomes and must be minimised wherever possible. Bias can be minimised by ensuring the research design is similar to the screen-reading process in practice, using criteria A1–A8 (Reference Standard; Disease Progression bias; Verification bias; Incorporation bias; Review bias; Clinical review bias; Instrument variation. A5 was not applicable). The highest number of negative responses for bias potential were A4 (Incorporation bias: 7.5), A8 (Observer/instrument variation: 11) and A9 (Handling of uninterpretable results: 11). Incorporation bias (A4) did occur in the studies since it is an immutable aspect of the screen-reading process. Potential confounders, which affect test performance and relate to the varying classification systems used (A8), can be reduced by using a validated reporting instrument such as the BIRADS® (Reston, Virginia) classification lexicon.26 Uninterpretable results (A9) were not included in the reviewed studies since the results were known prior to the test. Since potential bias is the predominant detractor of quality in the reviewed studies it is suggested that further work needs to identify the association of the criteria within category A with the estimates of diagnostic accuracy produced in the studies, and determine a hierarchy of the impact of negative responses to the criteria for outcome estimates of accuracy. Negative responses in section B: Applicability of results, were the highest in B4 (Change in technology of index test). This is explained by the introduction of digital technology since the reviewed articles were published. It is possible that this new technology may provide increased diagnostic accuracy in screen-reading, and so results of the reviewed studies may not be generalisable to facilities using the digital equipment. This change, however, did not influence the applicability of the results at the time of publication. Section C: Conduct of the study demonstrated low but significant negative responses to C2 (Sample size). Appropriate sample size is a critical component of a research study and in the field of research covered by the reviewed studies, sample refers to both number of images read and number of observers reading the images. This criterion, therefore, requires clarification. Section D: Reporting of the study criteria D6 (Precision of results: 8) and D7 (Drop-outs: 10) demonstrated large numbers of negative responses. Precision of results and accounting for all the images (rather than patients) were problematic in some studies. The way in which these criteria are expressed, however, does not readily apply to screen-reading. Section E: Observer characteristics demonstrated high numbers of negative responses in criteria E6 (Observer variability: 7) and E7 (Analysis of observer variability: 10) which are fundamentally the same. Observer variability should be analysed statistically through the use of the Kappa statistic or similar, as appropriate. In summary, these evaluation results emphasise the need for a specific evaluation tool for diagnostic accuracy in screen-reading. The specific screen-reading processes which minimise bias can be clearly enunciated, appropriate sample sizes of images and observers identified and criteria relating to study reporting increased in relevance.

Stage 4: Development of quality assessment tool for screen-reading studies

The development of topic-specific quality evaluation tools for diagnostic accuracy studies has been supported by Whiting et al.24 The quality tool used to evaluate the reviewed studies (Table1) was adapted to provide a specific tool for diagnostic accuracy studies in screen-reading and for ease of identification has been named the DASQUART (Diagnostic Accuracy Study Quality And Reporting Tool). The DASQUART is presented in Table4.
Table 4

Developed tool named DASQUART (Diagnostic Accuracy Study QUality And Reporting Tool) for determining quality in studies investigating diagnostic accuracy in screen-reading

CriteriaDescription of criteria
A1Reference standardAn appropriate reference standard of pathology and at least 1 year follow-up used to determine the presence or absence of breast cancer
A2Disease Progression biasAn interval cancer could not occur between the initial mammogram and the reference standard
A3Verification biasSame reference standard applied across the study
A4Incorporation biasThe reading of the screening mammogram does not form part of the reference standard
A6Review biasMammograms read blinded to knowledge of reference standard and interpretation by other readers
A7Clinical review biasPrevious image rounds available for comparison
A8Instrument variationNo reporting instrument variation which will affect estimates of test performance, e.g., use of BIRADS® lexicon26
A9Handling of uninterpretable resultsUninterpretable results included in the analysis
A10Arbitrary choice of threshold valueThreshold value of normal chosen independently of results
B1Spectrum compositionImage sample similar to one of interest (test sets, e.g., PERFORMS,31 BREAST32 and consecutive screening)
B2Population recruitmentImage sample selected adequate to include appropriate spectrum (test sets, e.g., PERFORMS,31 BREAST32)
B3Disease prevalence/severitySpectrum of breast cancer prevalence similar to one of interest (test sets, e.g., PERFORMS,31 BREAST32 and consecutive screening)
B4Change in technology of index testNo change in mammography technology which will affect applicability of results
C1Subgroup analysisSubgroup analyses were appropriate and specified
C2Sample sizeAppropriate number of images included in study
C3ObjectivesStudy objectives relevant to study question
C4Study designThe purpose, method, results and conclusions demonstrate logical coherence and consistency
D1Inclusion criteriaIncluded in systematic reviews
D2Test execution (a) imagesSufficient details of mammogram reading reported to permit its replication. Details include number of images read in total and at one sitting, how images were selected (test sets), degree of difficulty (test sets), types of breast cancers included (test sets).
D2Test execution (b) environmentTime taken to read, background lighting and type of monitors
D3Reference executionSufficient details provided of reference standard used to permit its replication
D4Normal definedAuthors clearly reported what was considered a normal reading result
D5Appropriate resultsAppropriate results of accuracy presented, e.g., sensitivity, specificity, ROC and JAFROC analysis
D6Precision of resultsEstimate of precision of results presented as appropriate
D7Drop-outsAll images and observers accounted for
D8Data tableTest performance reported in a data table
D9Utility of testClinical relevance of the test emphasised
E1Image allocation to observersImage allocation to observers described
E2Number of observersNumber of observers presented
E3Observer experienceExperience of observers described
E4Observer trainingTraining of observers described
E5Observer professionProfession of observers presented
E6Analysis of observer variabilityObserver variability in analysis, e.g., Kappa statistic
Developed tool named DASQUART (Diagnostic Accuracy Study QUality And Reporting Tool) for determining quality in studies investigating diagnostic accuracy in screen-reading The quality evaluation criteria of Whiting et al.17 and additional criteria related to medical imaging of Brealey and Westwood21 have been adapted to enhance relevance, clarity and precision and to contribute to the development of a user-friendly quality assessment tool. These adaptations are described below.

Changes to existing criteria

To maintain consistency in the structure of the tool, definitive statements rather than questions are presented throughout as descriptions of criteria. A positive response to these statements indicates an aspect of quality. Criteria for which a negative response indicates quality have been changed (A4: Incorporation bias, A8: Observer/instrument variation and B4: Change in technology of index test). One criterion not relevant to this area of study has been removed (A5: Treatment paradox) since treatment does not typically begin until verification has been made through pathology results. Criterion A8 of observer variation is similar to criteria E1–E7 of observer characteristics and has been removed. Only instrument variation, specifically the reporting form used to interpret the images, now comprises A8. For criterion C2 (Sample size), participants are changed to images while number of observers (screen-readers) comprises E2. The inclusion in D5 (Appropriate results) of receiver operating characteristic (ROC) and Jackknife Free-response Receiver Operating Characteristic (JAFROC) analysis rigorously assesses observer accuracy. This method allows quantitative analysis of observers interpreting images which could contain more than one lesion. For D7 (Drop-outs), patients are replaced by images and observers.

Additional criteria

Criterion D2 (Test execution) now provides further detail to allow replicability as well as identify variables which influence diagnostic accuracy to further adapt the tool to this area of study. Details included are related to the screen-reading process and include: number of images read at one sitting; how images were selected; degree of difficulty of interpretation; details of types of breast cancer; time taken to read; and environmental conditions such as lighting and type of monitors. Observer variability among radiologists has been found to be related to years of experience and numbers of images read22,23 and so these criteria have been added to E: Observer characteristics.

Evidence for criteria

This adaptation has been carried out using evidence from the literature: van den Biggelaar et al.16 (A1: Reference Standard, D5: Appropriate results), Reed et al.27 (D2: Test execution) and Brealey and Westwood21–23 (E1–E7: Observer Characteristics). As well, details of the breast screening process contained within the BreastScreen Australia National Accreditation Standards (NAS)1 were also used. The NAS is not only based on rigorous international evidence relating to best practice1 but also encourages the research design in these studies to mimic the real-life environment of screen-reading and consequently provide the most clinically useful outcomes. One aspect of the screen-reading process which is typically impractical for research purposes is screen-reading consecutive populations. This has led to the use of test sets in research studies. However, for these studies to be clinically useful a correlation between test set results and real-life clinical results is essential.

Test sets and clinical practice

Much debate surrounds the testing of diagnostic accuracy using test sets which have artificially inflated breast cancer prevalence versus consecutive screening images which mimic the real-life clinical situation. Minimal or no correlation between test set outcomes and clinical outcomes has been identified by Scott et al.,28 Rutter and Taplin.29 Gur et al. reported a significant difference between performance in the clinic than completing test sets.30 A study by Pauli et al. found a strong correlation between test set outcomes and consecutive screening outcomes when used together in the same research design.12 These studies, however, used varying numbers of breast cancers, images and types of breast cancer to comprise the test set. This variation can be overcome by the use of a validated test set such as PERFORMS (Scott and Gale)31 and BREAST (Brennan, Lee and Tapia)32 which increases the rigour of the study and provides consistency in the important aspects of study as spectrum composition, spectrum of images and spectrum of disease (B1–B3). The degree of difficulty in terms of types of cancers, proportions of breast density and numbers of images read would be consistent and so comparisons between study outcomes could be more readily applied. Incorporating validated test sets into the quality evaluation tool specifically developed to evaluate screen-reading accuracy, may well lead to an identification and understanding of the specific causal agents for any lack of correlation between clinical audits and screen-reading test sets, which as Soh et al. state is needed to facilitate the process of evaluating the diagnostic accuracy of screen-readers in practice.20

Validation of DASQUART

Since this reported analysis was carried out, an updated version of the QUADAS tool has been developed by Whiting et al. (plus other authors) and named Quadas-2 which includes a number of additional and improved features particularly focusing on bias.33 A review in 2013 by the same authors provides an updated classification and overview of the sources of bias and variation in test accuracy studies.34 As part of the validation process of the DASQUART careful analysis of these two Whiting et al. (plus other authors) studies should be undertaken to identify any further classifications and adaptations required to improve the tool.

Conclusion

This reported study has developed a quality assessment tool specifically for evaluating the quality of studies investigating the diagnostic accuracy of screen-readers. This tool now needs further refinement and rigorous validation processes including critical evaluation by a panel of clinical experts in the area of screen-reading. A limitation of this study is the focus on evaluation of studies only involving radiographers as screen-readers; future research should include a quality evaluation of studies conducted by radiologists as screen-readers to provide evidence for further refinement of the tool. This tool, with further refinement and validation, can make a significant contribution to promoting well-designed quality studies in this important area of research and practice. Most importantly it can facilitate consistency in study design which can increase the rigour and applicability of the outcomes to the clinical environment.

Conflict of Interest

The authors declare no conflict of interest.
  27 in total

1.  Assessing mammographers' accuracy. A comparison of clinical and test performance.

Authors:  C M Rutter; S Taplin
Journal:  J Clin Epidemiol       Date:  2000-05       Impact factor: 6.437

2.  An observational study to evaluate the performance of units using two radiographers to read screening mammograms.

Authors:  R L Bennett; S J Sellars; R G Blanks; S M Moss
Journal:  Clin Radiol       Date:  2011-11-08       Impact factor: 2.350

Review 3.  Assessing reader performance in radiology, an imperfect science: lessons from breast screening.

Authors:  B P Soh; W Lee; P L Kench; W M Reed; M F McEntee; A Poulos; P C Brennan
Journal:  Clin Radiol       Date:  2012-04-07       Impact factor: 2.350

4.  Breast screening: PERFORMS identifies key mammographic training needs.

Authors:  H J Scott; A G Gale
Journal:  Br J Radiol       Date:  2006-12       Impact factor: 3.039

5.  Are you reading what we are reading? The effect of who interprets medical images on estimates of diagnostic test accuracy in systematic reviews.

Authors:  S Brealey; M Westwood
Journal:  Br J Radiol       Date:  2007-08       Impact factor: 3.039

6.  Markers of good performance in mammography depend on number of annual readings.

Authors:  Mohammad A Rawashdeh; Warwick B Lee; Roger M Bourne; Elaine A Ryan; Mariusz W Pietrzyk; Warren M Reed; Robert C Heard; Deborah A Black; Patrick C Brennan
Journal:  Radiology       Date:  2013-06-04       Impact factor: 11.105

7.  Effects of a program to train radiologic technologists to identify abnormalities on mammograms.

Authors:  L W Bassett; A J Hollatz-Brown; R Bastani; J G Pearce; K Hirji; L Chen
Journal:  Radiology       Date:  1995-01       Impact factor: 11.105

8.  QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies.

Authors:  Penny F Whiting; Anne W S Rutjes; Marie E Westwood; Susan Mallett; Jonathan J Deeks; Johannes B Reitsma; Mariska M G Leeflang; Jonathan A C Sterne; Patrick M M Bossuyt
Journal:  Ann Intern Med       Date:  2011-10-18       Impact factor: 25.391

Review 9.  A systematic review classifies sources of bias and variation in diagnostic test accuracy studies.

Authors:  Penny F Whiting; Anne W S Rutjes; Marie E Westwood; Susan Mallett
Journal:  J Clin Epidemiol       Date:  2013-08-17       Impact factor: 6.437

10.  Inter-observer variability in mammography screening and effect of type and number of readers on screening outcome.

Authors:  L E M Duijm; M W J Louwman; J H Groenewoud; L V van de Poll-Franse; J Fracheboud; J W Coebergh
Journal:  Br J Cancer       Date:  2009-03-03       Impact factor: 7.640

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.