Literature DB >> 30345030

Audit of transvaginal sonography of normal postmenopausal ovaries by sonographers from the United Kingdom Collaborative Trial of Ovarian Cancer Screening (UKCTOCS).

Will Stott1, Aleksandra Gentry-Maharaj1, Andy Ryan1, Nazar Amso2, Mourad Seif3, Chris Jones1, Ian Jacobs1,4, Max Parmar5, Usha Menon1, Stuart Campbell6, Matthew Burnell1.   

Abstract

Background: We report on a unique audit of seven sonographers self-reporting high visualization rates of normal postmenopausal ovaries in the United Kingdom Collaborative Trial of Ovarian Cancer Screening (UKCTOCS).  This audit was ordered by the trial's Ultrasound Management Subcommittee after an initiative taken in 2008 to improve the quality of scanning and the subsequent increase in the number of sonographers claiming very high ovary visualisation rates.
Methods: Seven sonographers reporting high rates (>89%) of visualizing normal postmenopausal ovaries in examinations performed between 1 st January and 31 st December 2008 were identified. Eight experts in gynaecological scanning reviewed a random selection of exams performed by these sonographers and assessed whether visualization of both ovaries could be confirmed (cVR-Both) in the examinations. A random effects bivariate probit model was fitted to analyse the results.  
Results: The eight experts reviewed images from 357 examinations performed on 349 postmenopausal women (mean age 60.0 years, range 50.2-73.3) by the seven sonographers. The mean cVR-Both obtained from the model for these sonographers was 67.2% with a range of 47.6-86.5% (95%CI 63.9-70.5%). The range of cVR-Both between the experts was 47.3-88.3% and the intra-class correlation coefficient (ICC) for left and right ovary confirmation was 0.39.     Conclusions: The audit suggests that self-reported visualization of postmenopausal ovaries is unreliable, as visualisation of both ovaries could not be confirmed in almost a third of examinations. The agreement for visualization of both ovaries based on review of a static image between experts and sonographers and between expert reviewers alone was only moderate. Further research is needed to develop reliable Quality Control metrics for transvaginal ultrasound.

Entities:  

Keywords:  Audit; Ovarian Cancer Screening; Quality Control (QC); Transvaginal Sonography Scans (TVS); Ultrasound; Visualisation Rate (VR)

Mesh:

Year:  2018        PMID: 30345030      PMCID: PMC6173132          DOI: 10.12688/f1000research.15663.1

Source DB:  PubMed          Journal:  F1000Res        ISSN: 2046-1402


Introduction

The normal ovary of a postmenopausal woman is a small structure (mean volume 1.25ml [1]) usually situated lateral to the uterine fundus and in close relation to the internal iliac vein. In as many as 40% of transvaginal ultrasound (TVS) examinations [2] the ovary may not been seen as typically they shrink with age and are sometimes very difficult to locate [3, 4]. For this reason in the United Kingdom Collaborative Trial of Ovarian Cancer Screening (UKCTOCS) and other screening trials [2, 5, 6] a pragmatic approach is taken whereby an annual screening examination may be judged satisfactory even if both ovaries are not seen, given that a good view has been achieved of the Iliac vessels in the pelvic side wall. However, the sonographer should always attempt to visualize both ovaries as this provides the maximum assurance that an early ovarian cancer has been excluded. A metric commonly used in the quality control (QC) of TVS is self-reported visualisation rate (VR), defined as the number of examinations in which the ovaries were visualized as a proportion of all examinations performed by the sonographer [7]. In 2008, UKCTOCS implemented an accreditation programme which included the monitoring of individual sonographer VR over a 3 month period [8]. This revealed that some sonographers were self-reporting higher than expected VR. Therefore in 2009, it was decided to audit the performance of these high scoring sonographers to confirm independently whether it is possible to achieve high rates of ovary visualisation in postmenopausal women. We report on this audit and its outcome.

Methods

UKCTOCS trial

The TVS in this study were performed as part of the UKCTOCS, which is a multi-centre randomized controlled trial of 202,638 women volunteers from 13 trial centres throughout Northern Ireland, Wales and England ( ISRCTN22488978). The inclusion criteria specified by the trial protocol were postmenopausal women aged between 50–74 years. The women were randomised into three groups with the ultrasound arm involving 50,639 women who underwent annual TVS examinations. Sonographers performing the examinations were required to 1) record whether the ovary had been visualized, 2) measure the ovary in 3 orthogonal dimensions, and 3) comment on its morphology. These observations were stored centrally in the Trial Management System (TMS). The sonographer measured the dimensions of each ovary using digital callipers manually positioned on the extent of the ovary boundary in static images in two orthogonal planes during the examination; see Figure 1. The distance between the calliper marks was displayed in millimeters at the bottom of the image and copied into the TMS exam record fields as D1, D2 and D3. D1 represents the longest ovarian distance in longitudinal section (LS) and D2 is the widest distance (Anteroposterior - AP) which can be measured at 90° to the line used to measure D1. The largest diameter of the ovary in transverse section (TS) is measured as D3. These dimensions allow calculation of ovarian volume using the prolate ellipsoid formula; D1xD2xD3 x0.5423.
Figure 1.

Transverse (TS) and longitudinal (LS) transvaginal ultrasound images of left ovary acquired by sonographer.

This ovary was confirmed as normal and correctly measured by the expert reviewer.

Transverse (TS) and longitudinal (LS) transvaginal ultrasound images of left ovary acquired by sonographer.

This ovary was confirmed as normal and correctly measured by the expert reviewer. The TVS images used to measure the ovaries for each patient were saved on the ultrasound machines at each of the 13 trial centres and periodically copied onto disks which were sent by courier to the trial coordinating centre in London where they were copied into a bespoke computer system called the Ultrasound Record Archive (URA). These archived static images allow independent confirmation as to whether the feature measured was an ovary, thus permitting a subsequent audit of the sonographer’s self-reported VR.

Audit dataset

Sonographers who had performed >100 TVS exams between January 2008 and January 2009 and who had reported a high rate of ovary visualisation (>89%) over this period were identified. The audit dataset was created by assigning a random number to the annual exams performed by each of the sonographers during this same period and then making a random selection for each sonographer based on the value of these numbers. Inclusion criteria were both ovaries reported as visualized and the examination classified as having normal morphology. Examinations were excluded if the corresponding images were not stored in the URA. All exams audited were performed using a Medison Accuvix (model XQ, software v1.08.02, transvaginal probe type EC4-9IS 4-9 MHz).

Audit methodology

Eight members of the UKCTOCS Ultrasound Subcommittee who were highly experienced in gynaecological scanning undertook the review. They included three consultant gynaecologists, two gynaecological radiologists and three National Health Service (NHS) superintendent grade sonographers. Originally there were nine experts but it subsequently transpired that one of the reviewers was also one of the seven sonographers being audited. Therefore, it was decided to remove this reviewer’s results from the study. Accordingly, though these experts were initially split into three groups of three, one group was reduced to two experts following the exclusion of reviewer nine. The audit dataset was randomly split such that each group reviewed 119 exams (total 357 exams) and each expert was asked to assess 17 exams performed by each of the seven sonographers. In this way, each exam was judged by at least two separate experts. In order to avoid bias each expert was blinded as to the name of the sonographer being reviewed and the assessment of the other experts. The primary aim of the audit was to confirm the self-reported visualisation of both ovaries (cVR-Both) in examinations by each of the seven sonographers, which by extension required each expert reviewer to identify the exact images used to measure both ovaries from all of the images captured during the exam (mean 5.4, range 1–30). A software tool called osImageManager was developed specifically for the reviewers ( Figure 2). It facilitated display of the images associated with each of the examinations and also recorded the review results in the audit database.
Figure 2.

Screenshots of the osImageManager – the application used to facilitate Transvaginal Ultrasound Exam Review.

Statistical analysis

The baseline characteristics of the women are reported by trial centre code, age, years since last period, body mass index (BMI), hysterectomy status, oral contraceptive pill (OCP) and hormone replacement therapy (HRT) use. Information from the UKCTOCS sonographer accreditation records was used to calculate the mean, range and standard deviation of their collective experience. Their level of training and qualifications was also compared. Raw confirmed VR for each sonographer, each expert and overall were calculated for left ovary (LO) and right ovary (RO) as well as jointly for both LO and RO in the same examination. However, for formal inference we calculated the confirmed VR based on a statistical model.

Statistical modelling

All modelling was performed in Stata v14.2. The data was analysed using a bivariate probit random effects model. The bivariate outcome was the experts’ binary judgement of whether they confirmed the scan as seen or not seen, for both LO and RO. For the LO and RO portion of the model there was a scan-specific random intercept term representing the dependence of judgements within each scan, rated by three (or two) expert reviewers. The LO and RO random effects were allowed to covary as were the LO and RO error terms. In addition the model had categorical fixed effects for the original sonographer (n=7) and the expert (n=8). The details of the model can be found in Supplementary File 1. The model was fitted in Stata 14.2 with the user-written command cmp [9]. Two additional models were fitted. Firstly, one that included the factor ‘qualification’ (gynaecologist, radiologist, sonographer) instead of the factor ‘expert’ which, fully nested within ‘qualification’, meant both terms could not be included. Secondly, the factor ‘expert’ was simply taken out for reasons described in ‘Predictions and Correlations’. The use of this statistical model allowed us to simultaneously analyse all the data despite some scans being judged by a different number of experts. This included instances when only the LO or RO of a scan had been reviewed. By making use of model-based predictions, the model allowed us to assess the impact of each sonographer (or reviewer) whilst generalizing over the sample of reviewer (or sonographer) and volunteers, separately for LO and RO, but also for both ovaries in a joint manner. The raw proportions, summed over either sonographer or reviewer, fail to take in the within-volunteer correlation. All joint significance tests of the parameters were Wald tests. Stata’s post-estimation command margins were used to make predictions based on the probit model parameters. Specifically, marginal probability predictions were made over the whole sample, and for each sonographer and expert for both equations (LO and RO). In addition, the joint probability of a positive outcome for both LO and RO were calculated by incorporating the estimated correlation of both the random intercepts and error terms. All marginal predictions were ‘population-averaged’ in that they were integrated over the value range of the random effects. Individual random effects were calculated using empirical Bayes means. Separate intraclass correlation coefficients (ICC) for both LO and RO were calculated using the variance component estimates (see Supplementary File 1). The ICCs estimate the dependence between the dichotomous outcomes within the same volunteer, after taking into account the fixed effects. The ICC was also calculated based on a model with no ‘expert’ term, as its inclusion will provide an ICC that reflects within-scan correlation after adjusting for each expert’s general propensity to confirm visualisation. Supplementary File 1 also describes the calculation of the correlation between the left and right ovary result for a given volunteer on a given review occasion, necessary for the joint probability estimation. Note that the correlations from a probit model are ‘tetrachoric’ – that is, the correlation of two theorised normally distributed continuous latent variables, which produce the observed binary outcomes.

Results

An audit dataset of 357 annual TVS exams from 349 women was produced by making a random selection of 51 exams performed by each of the seven UKCTOCS sonographers who had reported ovary visualisation rates >89% for the exams they had performed during the study period (1/1/08 to 31/12/08) irrespective of outcome; normal, abnormal or unsatisfactory. However, only examinations with normal morphology reported were reviewed. Fifteen reviews were ineligible for various reasons. The eight expert reviewers performed the image review at locations in Derby, Manchester, Bristol and London. They collectively spent approximately 100 hours conducting their audit of the work of the seven UKCTOCS sonographers. The sonographers had a mean experience of 14.5 years (range 7–23, SD 7). They operated in five different trial centres with two pairs of sonographers working in the same centre. All sonographers were accredited by UKCTOCS during 2008. The 349 women whose exams were included in the audit dataset had a mean age of 60.0 years (range 50.2–73.3, SD 5.85), mean age at last period of 49.3 years (range 27.9–70.0, SD 5.66), mean BMI of 26.2 (range 17.5–45.1, SD 4.17), use of HRT at recruitment of 24.9%, ever use of OCP of 64.7% and a history of hysterectomy in 12.4%.

Model results

In total the model fitted 1871 ultrasound scan assessments formed from 940 LO scans and 931 RO scans resulting in 945 scans where at least one ovary was included. The fixed effects of both sonographer and expert were highly significant for either left or right ovary (joint p<0.0001 always, Table 1). As expected, the fitted predictions for LO or RO separately were close to the raw proportions over the same sample (see Table 2) because the design was (largely) balanced and the predictions did not include an adjusting variable. The overall LO prediction was 0.78 (95% CI: 0.75-0.81), but by sonographer this ranged from 0.65 to 0.89. By reviewer, the range was from 0.59 to 0.93. For RO, predicted probabilities were typically higher; overall prediction was 0.80 (95% CI: 0.77-0.83), sonographer predictions ranged from 0.62 to 0.97 and reviewer predictions ranged from 0.66 to 0.94. Not all sonographer or reviewer rank orderings were the same for LO and RO, for example reviewer 7 was the lowest for LO and reviewer 5 for RO. This was in contrast to the raw proportions where reviewer 7 gave the lowest percentage of confirmations for both LO and RO. In a separate model where expert was replaced by ‘qualification’, sonographers had significantly higher confirmed VR for both LO ( β=0.74 95% CI: 0.38-1.10) and RO ( β=0.86 95% CI: 0.40-1.32) compared to gynaecologists ( Table 1). Radiologists also had higher confirmed VR than gynaecologists but this was only significant at the 5% level for LO. The mean cVR-Both obtained using the model was 67.2%, ranging from 47.6% to 86.5% (95%CI: 63.9-70.5%, Table 2) and Figure 3 and Figure 4 present marginal joint predictions (cVR-Both) for individual experts and sonographers respectively.
Table 1.

Results of the random effects bivariate probit model – fixed and random effects.

Fixed effects
betastandard errorL95% CIU95% CIp-valueleft vs right
LEFT OVARY
sonographer ID (vs A)0.0000p=0.1153
Sonographer B0.5770.2740.0391.115
Sonographer C1.2020.3000.6151.789
Sonographer D0.1960.268-0.3300.722
Sonographer E1.1420.2950.5641.721
Sonographer F0.7730.2790.2251.320
Sonographer G0.0860.261-0.4250.597
reviewer ID (vs reviewer 1)0.0000p=0.7544
reviewer 20.6200.371-0.1061.347
reviewer 3-0.3540.313-0.9680.261
reviewer 4-0.2040.305-0.8020.394
reviewer 5-1.0470.301-1.636-0.457
reviewer 6-0.3620.316-0.9830.258
reviewer 7-1.1300.291-1.701-0.559
reviewer 8-0.1800.306-0.7810.421
Qualification (vs gynaecologist) * 0.0002p=0.313
sonographer0.7410.1830.3821.099
radiologist0.4420.1950.0600.825
constant0.8940.2820.3411.4470.0020
RIGHT OVARY
sonographer ID (vs A)0.4840.318-0.1401.107
Sonographer B1.6020.3690.8782.326
Sonographer C 0.7850.3310.1351.434
Sonographer D2.4700.4601.5693.371
Sonographer E1.1080.3420.4381.777
Sonographer F0.3600.308-0.2450.964
Sonographer G
reviewer ID (vs reviewer 1)0.0000
reviewer 20.5280.476-0.4051.462
reviewer 3-0.9220.392-1.691-0.154
reviewer 40.0200.398-0.7600.800
reviewer 5-1.3030.387-2.060-0.545
reviewer 6-0.5460.408-1.3470.254
reviewer 7-1.1820.367-1.901-0.464
reviewer 8-0.5010.383-1.2520.250
Qualification (vs gynaecologist) * 0.0010
sonographer0.8610.2360.3991.320
radiologist0.1330.233-0.3230.589
constant1.0030.3520.3131.6930.0040
Random effects and correlations p=0.4806
estimatestandard errorL95% CIU95% CI
left ovary RE variance0.7580.2170.3321.183p=0.210
right ovary RE variance1.2260.3300.5791.873
random effect covariance0.2930.1440.0110.576
random effect correlation0.3040.1260.0420.528
error term correlation0.4730.1070.2400.654
LO, RO correlation0.3870.0640.2620.513
left ovary ICC0.4310.0700.2940.569
right ovary ICC0.5510.0670.4200.681
left ovary ICC ** 0.3960.0680.2640.529
right ovary ICC ** 0.5070.0650.3790.635

*from a different model that replaces 'expert' with 'specialism'

** from a different model that excludes 'expert'

Table 2.

Results of the random effects bivariate probit model – marginal predictions.

POST ESTIMATION
Marginal predictions (population averaged)
probabilitystandard errorL95% CIU95% CIRaw proportion
LEFT OVARY=1
Overall0.7760.0150.7470.8060.774
sonographer A0.6530.0480.5580.7470.654
sonographer B0.7870.0410.7070.8670.780
sonographer C0.8920.0290.8350.9480.893
sonographer D0.7020.0470.6100.7930.729
sonographer E0.8840.0300.8250.9420.885
sonographer F0.8250.0360.7530.8960.817
sonographer G0.6750.0460.5840.7650.657
reviewer 10.8520.0310.7910.9130.847
reviewer 20.9320.0220.8880.9750.930
reviewer 30.7860.0370.7130.8590.793
reviewer 40.8160.0340.7480.8830.817
reviewer 50.6180.0440.5310.7040.632
reviewer 60.7840.0380.7100.8580.784
reviewer 70.5950.0440.5080.6810.585
reviewer 80.8200.0340.7530.8870.813
RIGHT OVARY=1
Overall0.8000.0150.7710.8300.799
sonographer A0.6240.0510.5230.7250.637
sonographer B0.7320.0470.6390.8250.731
sonographer C0.9060.0280.8510.9610.893
sonographer D0.7900.0430.7050.8740.807
sonographer E0.9690.0160.9371.0000.971
sonographer F0.8430.0370.7710.9150.825
sonographer G0.7060.0470.6140.7970.704
reviewer 10.8830.0280.8280.9380.890
reviewer 20.9350.0220.8920.9780.939
reviewer 30.7350.0400.6570.8130.750
reviewer 40.8850.0270.8330.9380.870
reviewer 50.6550.0420.5720.7380.675
reviewer 60.8040.0360.7330.8750.810
reviewer 70.6820.0400.6030.7600.644
reviewer 80.8120.0350.7440.8790.812
JOINT (LO=1, RO=1)
Overall0.6720.0170.6390.7050.670
sonographer A0.4760.0470.3840.5690.516
sonographer B0.6260.0470.5340.7180.618
sonographer C0.8280.0350.7600.8950.814
sonographer D0.6060.0470.5140.6980.621
sonographer E0.8650.0310.8040.9270.863
sonographer F0.7300.0410.6490.8120.714
sonographer G0.5380.0460.4480.6280.515
reviewer 10.7790.0350.7110.8470.780
reviewer 20.8830.0280.8280.9370.886
reviewer 30.6270.0400.5480.7060.612
reviewer 40.7510.0360.6810.8220.730
reviewer 50.4730.0400.3940.5520.470
reviewer 60.6740.0400.5950.7520.672
reviewer 70.4730.0400.3940.5510.492
reviewer 80.7040.0380.6300.7790.723
Figure 3.

Variation in cVR-Both by individual experts as calculated by the random effects bivariate probit model.

Figure 4.

Variation in cVR-Both by individual sonographers as calculated by the random effects bivariate probit model.

*from a different model that replaces 'expert' with 'specialism' ** from a different model that excludes 'expert' The variance estimates for the LO and RO random effects were 0.76 and 1.23 respectively ( Table 1), but this did not differ statistically (p=0.210). Indeed, despite the observed differences, there was no statistical difference in the LO versus RO effects concerning sonographer (p=0.115), reviewer (p=0.754) or the model as whole (p=0.481). The correlation of the LO and RO ovary random effects was 0.30 (95% CI: 0.04-0.53) and the error term correlation was 0.47 (95% CI: 0.24-0.65), implying a correlation of 0.39 (95% CI: 0.26-0.51) for the paired outcome of LO and RO for a given volunteer and occasion. This compares to the tetrachoric correlation of raw data of 0.51, and to 0.37 when the fixed effects are included in a standard bivariate probit model. The resultant within-volunteer correlation (ICC) for the repeated outcomes within a volunteer were 0.43 (95% CI: 0.29-0.57) and 0.55 (95% CI: 0.42-0.68) for LO and RO respectively. In addition, the ICCs for a model excluding the mean effect of the ‘expert’ term, were lower at 0.40 (95% CI: 0.26-0.53) for LO and 0.51 (95% CI: 0.38-0.64) for RO. Click here for additional data file.

Discussion

Our audit suggests that sonographer’s self-reported visualization rates of postmenopausal ovaries they judged to have normal morphology is unreliable. Our study was facilitated by the unique TMS and URA systems employed in UKCTOCS which permitted a retrospective review of the images and measurements recorded by the sonographer. It could be argued that the static images used for this audit represent a snapshot of a continuous pelvis examination so might not truly represent what was seen by the sonographer. Nevertheless, these static images were used to measure the ovaries, so the structure marked by the callipers was definitely considered to be an ovary by the sonographer. We analysed the data using a statistical model that accounted for the correlated structure of the data, between left and right ovary scans, and between the same scan viewed by the experts. Normality was assumed for the underlying latent variable (‘propensity to confirm visualisation’) and for the distribution of the ovary-specific volunteer random effects. The model gave predictions in the probability scale that different only slightly from the raw proportions, due to the nature of the study design. One clear benefit to using a statistical model with random effects is that all the data could be analysed together, and producing variance component estimates that allow the calculation of ICCs. The value of the ICC was higher for the right ovary then left, though not significantly different, and for both were modest: 0.40 for LO and 0.51 for RO when excluding the expert term from the fixed effects, the only variable that varied over each scan’s repeated assessments. Hence the ICC is a measure of inter-rater (expert) agreement, and suggests that although there is moderate concordance, the experts cannot be relied upon to replicate the judgements of each other. However, such lack of agreement in respect of each individual scan does not change of the overall conclusion of the audit in terms of the unreliability of the sonographer’s self-reported visualization rates. We have previously reported on the Quality Control (QC) of UKCTOCS TVS scanning with similar exam selection criteria (ovaries were seen and normal) [7]. A single expert reviewed 1000 randomly chosen TVS examinations which had been performed by 96 sonographers. The expert’s cVR-Both was 50% compared to the 100% VR as self-reported by the sonographers for these examinations. This result is broadly consistent with the results reported in this study for the group of seven sonographers with mean cVR-Both of 67.2%. The significant variation in cVR-Both across sonographer of normal postmenopausal ovaries is probably due to differences in sonographer ability and the subjective nature of this examination; a supposition supported by findings reported by Sharma et al. [8].

Limitations of the study

Intra-observer reproducibility was not addressed so the capability of individual experts to provide consistent results for the same exams was not measured. The study design was generally balanced, and potential confounders that might possibly affect visualization should be expected to be evenly distributed across experts due to the randomization process. However, it is conceivable that these confounders may not be balanced across sonographers, due to potential geographical differences in their distribution. This was not a major concern, but the factors could have been seamlessly absorbed into the model and produced sonographer predictions conditional on equal covariate distribution.

Conclusion

The results of this audit confirm that the visualization of postmenopausal normal ovaries by seven ‘high performing’ sonographers, as assessed by eight experts, could not be considered reliable given that in almost a third of their examinations structures other than an ovary had been mistakenly measured in at least one of the ovaries. However, individual sonographer performance varied significantly from 47% to 87% cVR-Both. These results show that it is possible for some sonographers to correctly visualize both ovaries when scanning a range of menopausal women so raising the possibly that other sonographers might achieve similar results if supported by a suitable quality improvement programme. This audit highlights the problem of sonographers routinely mistaking other features like the bowel as ovaries when scanning postmenopausal women. It also highlights the difficulties of providing effective Quality Control (QC) for such scans in a large scale screening programme. Specifically, it shows that undertaking the type of expert review conducted by this study for a substantial number of sonographers on a regular basis would not be feasible without creating dedicated teams specializing in normal ovary identification from TVS images of postmenopausal women. Therefore there is a need for further research to explore how independent and reliable QC metrics for TVS might be obtained by other means, for example by the automated analysis of TVS scan images both static and video. Recent advances in machine learning research, particularly in the area of deep neural networks, suggest it might soon be viable to construct a system able to determine sonographer VR from a collection of images captured during a series of TVS examinations. Indeed, the use of such deep learning techniques in the gathering of quality metrics from obstetric ultrasound images is already reporting some promise [10]. The work done by the UKCTOCS group on the QC of TVS scanning seeks to improve understanding of challenges associated with performing screening for ovarian cancer on a large scale and at multiple centres. All previous studies of ultrasound screening of postmenopausal ovaries for the early detection of cancer (excepting the recent QC study by our group) have accepted the self-reporting of ovarian visualisation rates as accurate. This is the first published audit of self-reporting of ovarian visualization rates and the results cause us to question the reliability of this metric, particularly for QC purposes.

Ethics approval

The UKCTOCS study was approved by North West Multicentre Research Ethics Committee 21/6/2000; MREC reference 00/8/34. It is registered as an International Standard Randomised Controlled Trial (no. ISRCTN22488978).

Data availability

The data referenced by this article are under copyright with the following copyright statement: Copyright: © 2018 Stott W et al. Data associated with the article are available under the terms of the Creative Commons Zero "No rights reserved" data waiver (CC0 1.0 Public domain dedication). Dataset 1: DataKey.txt – description of data fields; UKCTOCS TVC audit data biprobit format-0.csv; UKCTOCS TVC audit data biprobit format-0.dta; UKCTOCS TVC audit data do file.do. DOI, 10.5256/f1000research.15663.d213048 [11] Stata v14.2 was used in conjunction with the files in Dataset 1 to obtain the results presented in this paper. The United Kingdom Collaborative Trial of Ovarian Cancer Screening was a high-quality randomized clinical trial evaluating trans-vaginal ultrasound screening (TVUS) compared to a multi-modal screening (MMS) with Risk of Ovarian Cancer Algorithm (ROCA) testing based on CA-125 measurements followed by ultrasound when indicated compared to no screening in 202,546 women of average risk of ovarian cancer.  After 7 to 11 rounds of screening and up to 14 years of follow-up results trended toward benefit with ultrasound (HR = 0.91 [95% CI, 0.76 to 1.09]) and MMS (HR = 0.89 [95% CI, 0.74 to 1.08]) [1]. A comprehensive quality assurance program for TVUS was undertaken [2]. This current paper is the result of an audit ordered by the trial’s Ultrasound Management Subcommittee of seven sonographers reporting rates of visualizing both ovaries of > 89% after an accreditation program done by UKCTOCS in 2008.  Eight experts reviewed 357 archived, static images centrally stored that also had measurement markers recording ovary dimensions.  The mean visualization rate upon review for both ovaries fell to 67.2% with a range of 47.6 % to 86.5%.  The range between the experts was 47.3% to 88.3%. The trialists are to be commended for the design and conduct of this impressive ovarian cancer screening trial which shows some evidence of stage-shift and mortality reduction but after many rounds and long follow-up in a large group of women.  The quality assurance plans in place to train and monitor TVUS results is impressive.  While the technology dates to 2008-2009 and highly-selected experts might do better, the results of this audit strike a cautionary note about the rate of probably unavoidable non-visualization of ovaries with TVUS in post-menopausal women.  They call for expanded efforts in automated analysis with machine learning.  Improved technologies or biomarkers are also needed to realize the promise of lowering ovarian cancer mortality with early detection. I have read this submission. I believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard. This manuscript assesses the retrospective findings of 8 expert reviewers in assessing ultrasounds reported by 7 ultrasonographers who recorded visualization of both normal postmenopausal ovaries in >89% of scans in the United Kingdom Collaborative Trial of Ovarian Cancer Screening (UKCTOCS). The study set included 357 examinations from 349 women and 1,871 randomly chosen scans reviewed by these ultrasonographers selected from a one-year period in the trial. The main aim was to assess whether the interpretation of bilateral visualization of normal ovaries is reliable, comparing original reports to retrospective expert review. Data were analyzed using a bivariate probit random effects model with outcomes of bilateral visualization versus not. Notably, normal postmenopausal ovaries may be quite small and the average age of women in UKCTOCS was about 60 years. The data show poor reliability across measures: bilateral visualization for original reviewers showed a mean of 67%, with a range of 47.6%-86.5%; expert results ranged from 47.3%-88.3%. Agreement among expert reviewers was also modest. The authors conclude that further research is needed to develop reliable quality control metrics for transvaginal ultrasound. It is unclear how much expert reviewers knew about methods for assembling the study set and other details, which may have influenced interpretations. The study did not include a random sample of scans from the trial to serve as “distractors” or to provide a reference for comparison, and reliability data were not compared with external standards, such as measurements of ovaries that may have been removed later, CA 125 levels or rare cancer outcomes. External reviewers who were uninvolved with UKCTOCS may be of interest and possibly could be achieved via a web-based approach, at least for a subset. This study in combination with prior reports from UKCTOCS (e.g. Stott et al and Sharma et al) provides a composite picture of the performance of ultrasound in the trial; however, the generalizability of the current study is difficult to assess, given the unusual method of scan selection and the engagement of reviewers who were intimately knowledgeable about the trial and perhaps aware of the design of this project. Irrespective of these concerns, the data from UKCTOCS suggest that ultrasound of normal ovaries among older women has limitations. Given that reviewers were experienced and specifically trained for the task at hand, there are additional unknowns about whether and how performance could be improved, and how much  of reviewers’ performances reflect inherent limitations of ultrasound for assessment of ovaries and ovarian cancer screening.  Bodelon et al [1] reported a high frequency of non-visualization of ovaries in the Prostate, Lung, Colorectal, and Ovarian (PLCO) screening trial, with a tendency for individual women to have repeated non-visualization. Further, although non-visualization is likely a marker of smaller ovarian size on average, it is notable that non-visualization conferred at best a marginally reduced risk of developing ovarian cancer in PLCO. Analysis of serial ovarian volumes in PLCO suggested that enlargement occurs rapidly within one to two years of cancer detection, and therefore, would be unlikely to have meaningful impact on clinical outcomes. In a narrow sense, if ultrasound is to be used for ovarian cancer screening, then a better quality control metric than the frequency with which ovaries are visualized is needed. In a broader sense, this study and related literature call into question whether ultrasound imaging is useful in ovarian cancer screening, especially for high-grade serous carcinomas. To date, ovarian cancer screening and CA-125 has failed to achieve a reduction in ovarian cancer mortality. Although unproven, growing evidence points to the origin of many high-grade serous carcinomas, the most frequent lethal type of ovarian cancer, from the distal fallopian tube (fimbria), rather than from the ovarian surface epithelium. In contrast, other ovarian cancers (i.e. endometrioid and clear cell) may arise from endometriosis in the ovary and tend to remain organ confined for lengthier periods (present as stage I). Animal models of tubal cancer have shown that spread to the ovaries may accelerate disease progression (Perets et al [2]), but many questions about the pathogenesis of serous cancers among women, including the sojourn time of disease development and the role of the ovary in promoting metastatic spread. These fundamental questions raise larger issues about the role of assessing the ovary as part of cancer screening and the potential of ultrasound to identify cancers at early curable stage. Greater knowledge of the pathogenesis of ovarian cancer, especially high-grade serous cancers, may pose a barrier to improved early detection. Larger issues could be addressed to place the results of the study in context. I have read this submission. I believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however I have significant reservations, as outlined above. We thank Prof Sherman for his review, but seek clarification so that we might improve our paper. We note that he does not consider our conclusions are adequately supported by our results. Does he believe our results show that than the combined judgement of a team of eight experts when the reviewing the images the sonographer used to measure ovaries as taken from a random sample her scans over a year? If so, we should be grateful if he would identify the data in our results that supports such a conclusion. This paper reports the audit results of seven sonographer self-reporting high visualization rates of normal postmenopausal ovaries in the United Kingdom Collaborative Trial of Ovarian Cancer Screening (UKCTOCS). Eight experts reviewed static images from 357 ultrasound examinations performed on 349 postmenopausal women (mean age 60.0 yrs) to assess whether visualization of both ovaries with normal morphology could be confirmed. A random effects bivariate probit model was fitted to analyze the results. Both normal ovaries could be visualized in only two-thirds of cases, in both the sonographer and expert groups, and there was variation between findings reported by sonographers and their expert reviewers. As a result, the authors conclude that self-reported visualization of normal ovaries by sonographers is unreliable. The authors further suggest that effective quality control in the interpretation of large numbers of ovarian ultrasound images is difficult, and may be enhanced by applying recent advances in machine-learning to this problem. This audit confirms that the ultrasound evaluation of normal postmenopausal ovaries is often challenging because of their low volume. In fact, ovarian volumes continue to decrease after menopause, so ultrasound visualization may be even more difficult in women in the seventh and eighth decades of life [1]. As demonstrated, interpretation of ovarian ultrasound images is subjective, and may vary significantly between sonographers and expert physicians. For this reason, ultrasound scans, both of normal and abnormal ovaries, are reviewed by physicians for all women enrolled in the University of Kentucky Ovarian Screening Trial [2]. This review is time-consuming, however, and may be impractical as ovarian cancer screening is made available to large populations of at-risk women. The authors are to be commended for supporting efforts to adapt machine-learning technology to achieve automated analysis of static and video ultrasound images. A related and equally important issue is the ability of sonographers and expert reviewers to identify those morphologic patterns associated with early ovarian cancer. The evaluation algorithm used to identify those ultrasound patterns associated with ovarian cancer is important because it increases the positive predictive value of screening, and limits operative intervention for benign disease. We agree with the authors that ovarian cancer screening should be offered to women at moderate risk for ovarian cancer as a method to reduce disease mortality through early detection [3], and applaud their efforts to standardize ultrasound interpretation. We have read this submission. We believe that we have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.
  13 in total

1.  Ovarian volume related to age.

Authors:  E J Pavlik; P D DePriest; H H Gallion; F R Ueland; M B Reedy; R J Kryscio; J R van Nagell
Journal:  Gynecol Oncol       Date:  2000-06       Impact factor: 5.482

2.  Transformation of the fallopian tube secretory epithelium leads to high-grade serous ovarian cancer in Brca;Tp53;Pten models.

Authors:  Ruth Perets; Gregory A Wyant; Katherine W Muto; Jonathan G Bijron; Barish B Poole; Kenneth T Chin; Jin Yun H Chen; Anders W Ohman; Corey D Stepule; Soongu Kwak; Alison M Karst; Michelle S Hirsch; Sunita R Setlur; Christopher P Crum; Daniela M Dinulescu; Ronny Drapkin
Journal:  Cancer Cell       Date:  2013-12-09       Impact factor: 31.743

3.  Ovarian volume: determinants and associations with cancer among postmenopausal women.

Authors:  Mark E Sherman; James V Lacey; Saundra S Buys; Douglas J Reding; Christine D Berg; Craig Williams; Patricia Hartge
Journal:  Cancer Epidemiol Biomarkers Prev       Date:  2006-08       Impact factor: 4.254

4.  Long-term survival of women with epithelial ovarian cancer detected by ultrasonographic screening.

Authors:  John Rensselaer van Nagell; Rachel Ware Miller; Christopher P DeSimone; Frederick R Ueland; Iwona Podzielinski; Scott T Goodrich; Jeff W Elder; Bin Huang; Richard J Kryscio; Edward John Pavlik
Journal:  Obstet Gynecol       Date:  2011-12       Impact factor: 7.661

5.  Sensitivity and specificity of multimodal and ultrasound screening for ovarian cancer, and stage distribution of detected cancers: results of the prevalence screen of the UK Collaborative Trial of Ovarian Cancer Screening (UKCTOCS).

Authors:  Usha Menon; Aleksandra Gentry-Maharaj; Rachel Hallett; Andy Ryan; Matthew Burnell; Aarti Sharma; Sara Lewis; Susan Davies; Susan Philpott; Alberto Lopes; Keith Godfrey; David Oram; Jonathan Herod; Karin Williamson; Mourad W Seif; Ian Scott; Tim Mould; Robert Woolas; John Murdoch; Stephen Dobbs; Nazar N Amso; Simon Leeson; Derek Cruickshank; Alistair McGuire; Stuart Campbell; Lesley Fallowfield; Naveena Singh; Anne Dawnay; Steven J Skates; Mahesh Parmar; Ian Jacobs
Journal:  Lancet Oncol       Date:  2009-03-11       Impact factor: 41.316

6.  Sonographic visualization of the normal postmenopausal ovary.

Authors:  D A Hall; K A McCarthy; D B Kopans
Journal:  J Ultrasound Med       Date:  1986-01       Impact factor: 2.153

7.  Sonographers' self-reported visualization of normal postmenopausal ovaries on transvaginal ultrasound is not reliable: results of expert review of archived images from UKCTOCS.

Authors:  W Stott; S Campbell; A Franchini; O Blyuss; A Zaikin; A Ryan; C Jones; A Gentry-Maharaj; G Fletcher; J Kalsi; S Skates; M Parmar; N Amso; I Jacobs; U Menon
Journal:  Ultrasound Obstet Gynecol       Date:  2018-03       Impact factor: 7.299

8.  Factors affecting visualization of postmenopausal ovaries: descriptive study from the multicenter United Kingdom Collaborative Trial of Ovarian Cancer Screening (UKCTOCS).

Authors:  A Sharma; M Burnell; A Gentry-Maharaj; S Campbell; N N Amso; M W Seif; G Fletcher; C Brunel; G Turner; R Rangar; A Ryan; I Jacobs; U Menon
Journal:  Ultrasound Obstet Gynecol       Date:  2013-10       Impact factor: 7.299

9.  Quality assurance and its impact on ovarian visualization rates in the multicenter United Kingdom Collaborative Trial of Ovarian Cancer Screening (UKCTOCS).

Authors:  A Sharma; M Burnell; A Gentry-Maharaj; S Campbell; N N Amso; M W Seif; G Fletcher; C Brunell; G Turner; R Rangar; A Ryan; I Jacobs; U Menon
Journal:  Ultrasound Obstet Gynecol       Date:  2016-02       Impact factor: 7.299

10.  Ovarian cancer screening and mortality in the UK Collaborative Trial of Ovarian Cancer Screening (UKCTOCS): a randomised controlled trial.

Authors:  Ian J Jacobs; Usha Menon; Andy Ryan; Aleksandra Gentry-Maharaj; Matthew Burnell; Jatinderpal K Kalsi; Nazar N Amso; Sophia Apostolidou; Elizabeth Benjamin; Derek Cruickshank; Danielle N Crump; Susan K Davies; Anne Dawnay; Stephen Dobbs; Gwendolen Fletcher; Jeremy Ford; Keith Godfrey; Richard Gunu; Mariam Habib; Rachel Hallett; Jonathan Herod; Howard Jenkins; Chloe Karpinskyj; Simon Leeson; Sara J Lewis; William R Liston; Alberto Lopes; Tim Mould; John Murdoch; David Oram; Dustin J Rabideau; Karina Reynolds; Ian Scott; Mourad W Seif; Aarti Sharma; Naveena Singh; Julie Taylor; Fiona Warburton; Martin Widschwendter; Karin Williamson; Robert Woolas; Lesley Fallowfield; Alistair J McGuire; Stuart Campbell; Mahesh Parmar; Steven J Skates
Journal:  Lancet       Date:  2015-12-17       Impact factor: 79.321

View more
  1 in total

1.  Performance Characteristics of the Ultrasound Strategy during Incidence Screening in the UK Collaborative Trial of Ovarian Cancer Screening (UKCTOCS).

Authors:  Jatinderpal Kalsi; Aleksandra Gentry-Maharaj; Andy Ryan; Naveena Singh; Matthew Burnell; Susan Massingham; Sophia Apostolidou; Aarti Sharma; Karin Williamson; Mourad Seif; Tim Mould; Robert Woolas; Stephen Dobbs; Simon Leeson; Lesley Fallowfield; Steven J Skates; Mahesh Parmar; Stuart Campbell; Ian Jacobs; Alistair McGuire; Usha Menon
Journal:  Cancers (Basel)       Date:  2021-02-18       Impact factor: 6.639

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.